Proceedings of the Twenty-Fifth International Joint Conference on Artificial Intelligence, IJCAI 2016, New York, NY, USA, 9-15 July 2016. IJCAI/AAAI Press 【DBLP Link】
【Paper Link】 【Pages】:xxxiii-xxxiv
【Authors】: Subbarao Kambhampati ; Gerhard Brewka
【Abstract】: IJCAI-2016 is different from earlier conferences in the IJCAI series in various respects. It is the first conference of its kind taking place one year after the last IJCAI (and the first-ever leap year one!). In our view one of the major roles of the flagship AI conference is to provide a forum for reintegrating diverse subfields of AI. For this reason we decided not to have vertical special tracks this year: that is, special tracks divided along specific subtopics such as knowledge representation, machine learning, planning and the like. We did introduce a horizontal special track, though, namely the AI and Web track. This track was chaired by Evgeniy Gabrilovich and Mausam and represents one of the most exciting current application areas for a broad range of AI techniques.
【Keywords】:
【Paper Link】 【Pages】:xxxv-xxvi
【Authors】:
【Abstract】: List of persons responsible for organizing and putting on the 2016 IJCAI conference.
【Keywords】:
【Paper Link】 【Pages】:xxxvii-lii
【Authors】:
【Abstract】: The 2016 program committees for the IJCAI main and special tracks.
【Keywords】:
【Paper Link】 【Pages】:liii-liv
【Authors】:
【Abstract】: Sponsors and organizers of the 2016 IJCAI conference.
【Keywords】:
【Paper Link】 【Pages】:lv-lvi
【Authors】:
【Abstract】: List of the 2016 recipients of the Award for Research Excellence, the John McCarthy Award, the Computers and Thought Award, the Donald E. Walker Distinguished Service Award, and the Distinguished Papers Award.
【Keywords】:
【Paper Link】 【Pages】:lvii-liix
【Authors】:
【Abstract】: List of persons in the IJCAI organization.
【Keywords】:
【Paper Link】 【Pages】:lix
【Authors】:
【Abstract】: List of all previous IJCAI conferences.
【Keywords】:
【Paper Link】 【Pages】:2-9
【Authors】: Andrés Abeliuk ; Haris Aziz ; Gerardo Berbeglia ; Serge Gaspers ; Petr Kalina ; Nicholas Mattei ; Dominik Peters ; Paul Stursberg ; Pascal Van Hentenryck ; Toby Walsh
【Abstract】: We propose a model of interdependent scheduling games in which each player controls a set of services that they schedule independently. A player is free to schedule his own services at any time; however, each of these services only begins to accrue reward for the player when all predecessor services, which may or may not be controlled by the same player, have been activated. This model, where players have interdependent services, is motivated by the problems faced in planning and coordinating large-scale infrastructures, e.g., restoring electricity and gas to residents after a natural disaster or providing medical care in a crisis when different agencies are responsible for the delivery of staff, equipment, and medicine. We undertake a game-theoretic analysis of this setting and in particular consider the issues of welfare maximization, computing best responses, Nash dynamics, and existence and computation of Nash equilibria.
【Keywords】:
【Paper Link】 【Pages】:10-16
【Authors】: Pritee Agrawal ; Pradeep Varakantham ; William Yeoh
【Abstract】: Synergistic interactions between task/resource allocation and stochastic planning exist in many environments such as transportation and logistics, UAV task assignment and disaster rescue. Existing research in exploiting these synergistic interactions between the two problems have either only considered domains where tasks/resources are completely independent of each other or have focussed on approaches with limited scalability. In this paper, we address these two limitations by introducing a generic model for task/resource constrained multi-agent stochastic planning, referred to as TasC-MDPs. We provide two scalable greedy algorithms, one of which provides posterior quality guarantees. Finally, we illustrate the high scalability and solution performance of our approaches in comparison with existing work on two benchmark problems from the literature.
【Keywords】:
【Paper Link】 【Pages】:17-23
【Authors】: Nirav Ajmeri ; Jiaming Jiang ; Rada Chirkova ; Jon Doyle ; Munindar P. Singh
【Abstract】: To interact effectively, agents must enter into commitments. What should an agent do when these commitments conflict? We describe Coco, an approach for reasoning about which specific commitments apply to specific parties in light of general types of commitments, specific circumstances, and dominance relations among specific commitments. Coco adapts answer-set programming to identify a maximal set of nondominated commitments. It provides a modeling language and tool geared to support practical applications.
【Keywords】:
【Paper Link】 【Pages】:24-30
【Authors】: Natasha Alechina ; Mehdi Dastani ; Brian Logan
【Abstract】: We consider the problem of whether a coalition of agents has a knowledge-based strategy to ensure some outcome under a resource bound. We extend previous work on verification of multi-agent systems where actions of agents produce and consume resources, by adding epistemic pre- and postconditions to actions. This allows us to model scenarios where agents perform both actions which change the world, and actions which change their knowledge about the world, such as observation and communication. To avoid logical omniscience and obtain a compact model of the system, our model of agents' knowledge is syntactic. We define a class of coalition-uniform strategies with respect to any (decidable) notion of coalition knowledge. We show that the model-checking problem for the resulting logic is decidable for any notion of coalition-uniform strategies in these classes.
【Keywords】:
【Paper Link】 【Pages】:31-37
【Authors】: Georgios Amanatidis ; Georgios Birmpas ; Evangelos Markakis
【Abstract】: We study a fair division problem with indivisible items, namely the computation of maximin share allocations. Given a set of n players, the maximin share of a single player is the best she can guarantee to herself, if she would partition the items in any way she prefers, into n bundles, and then receive her least desirable bundle. The objective then is to find an allocation, so that each player is guaranteed her maximin share. Previous works have studied this problem purely algorithmically, providing constant factor approximation algorithms. In this work, we embark on a mechanism design approach and investigate the existence of truthful mechanisms. We propose three models regarding the information that the mechanism attempts to elicit from the players, based on the cardinal and ordinal representation of preferences. We establish positive and negative (impossibility) results for each model and highlight the limitations imposed by truthfulness on the approximability of the problem. Finally, we pay particular attention to the case of two players, which already leads to challenging questions.
【Keywords】:
【Paper Link】 【Pages】:38-45
【Authors】: Giovanni Amendola ; Gianluigi Greco ; Nicola Leone ; Pierfrancesco Veltri
【Abstract】: A compact representation for non-transferable utility games founding on answer set programming is proposed. The representation is fully expressive, in that it can capture all games defined over a finite set of alternatives. Moreover, due to the knowledge representation capabilities of answer set programs, it can easily accommodate the definition of games within a wide range of application domains, ranging from scheduling, to routing and planning, just to name a few. The computational complexity of the proposed framework is studied, in particular, by focusing on the core as the prototypical solution concept. A system supporting the basic reasoning tasks arising therein is also made available, and results of experimental activity are discussed.
【Keywords】:
【Paper Link】 【Pages】:46-59
【Authors】: Elliot Anshelevich ; John Postl
【Abstract】: We determine the quality of randomized social choice mechanisms in a setting in which the agents have metric preferences: every agent has a cost for each alternative, and these costs form a metric. We assume that these costs are unknown to the mechanisms (and possibly even to the agents themselves), which means we cannot simply select the optimal alternative, i.e. the alternative that minimizes the total agent cost (or median agent cost). However, we do assume that the agents know their ordinal preferences that are induced by the metric space. We examine randomized social choice functions that require only this ordinal information and select an alternative that is good in expectation with respect to the costs from the metric. To quantify how good a randomized social choice function is, we bound the distortion, which is the worst-case ratio between expected cost of the alternative selected and the cost of the optimal alternative. We provide new distortion bounds for a variety of randomized mechanisms, for both general metrics and for important special cases. Our results show a sizable improvement in distortion over deterministic mechanisms.
【Keywords】:
【Paper Link】 【Pages】:53-59
【Authors】: Vincenzo Auletta ; Ioannis Caragiannis ; Diodato Ferraioli ; Clemente Galdi ; Giuseppe Persiano
【Abstract】: Recently, much attention has been devoted to discrete preference games to model the formation of opinions in social networks. More specifically, these games model the agents' strategic decision of expressing publicly an opinion, which is a result of an interplay between the agent's private belief and the social pressure. However, these games have very limited expressive power; they can model only very simple social relations and they assume that all the agents respond to social pressure in the same way. In this paper, we define and study the novel class of generalized discrete preference games. These games have additional characteristics that can model social relations to allies or competitors and complex relations among more than two agents. Moreover, they introduce different levels of strength for each relation, and they personalize the dependence of each agent to her neighborhood.
【Keywords】:
【Paper Link】 【Pages】:60-66
【Authors】: Haris Aziz ; Jérôme Lang ; Jérôme Monnot
【Abstract】: Selecting a set of alternatives based on the preferences of agents is an important problem in committee selection and beyond. Among the various criteria put forth for desirability of a committee, Pareto optimality is a minimal and important requirement.As asking agents to specify their preferences over exponentially many subsets of alternatives is practically infeasible,we assume that each agent specifies a weak order on single alternatives, from which a preference relation over subsets is derived using some preference extension.We consider four prominent extensions (responsive, leximax, best, and worst). For each of them, we consider the corresponding Pareto optimality notion, and we study the complexity of computing and verifying Pareto optimal outcomes. We also consider strategic issues: for three of the set extensions, we present linear-time, Pareto optimal and strategyproof algorithms that work even for weak preferences.
【Keywords】:
【Paper Link】 【Pages】:67-73
【Authors】: Haris Aziz ; Ildikó Schlotter ; Toby Walsh
【Abstract】: We initiate the study of control actions in fair division problems where a benevolent or malicious central organizer changes the structure of the fair division problem for self-interest or to benefit one, some or all agents. One motivation for such control is to improve fairness by minimally changing the problem. As a case study, we consider the problem of adding or deleting a small number of items to improve fairness. For two agents, we present polynomial-time algorithms for adding or deleting the minimum number of items to achieve ordinal envy-freeness. For three agents, we show that both problems, as well as the more basic problem of checking whether an envy-free allocation exists, are NP-complete. This closes a problem open for over five years. Our framework leads to a number of interesting directions in the area of fair division.
【Keywords】:
【Paper Link】 【Pages】:74-80
【Authors】: Yoram Bachrach ; Yuval Filmus ; Joel Oren ; Yair Zick
【Abstract】: Weighted voting games model decision-making bodies where decisions are made by a majority vote. In such games, each agent has a weight, and a coalition of agents wins the game if the sum of the weights of its members exceeds a certain quota. The Shapley value is used as an index for the true power held by the agents in such games. Earlier work has studied the implications of setting the value of the quota on the agents' power under the assumption that the game is given with a fixed set of agent weights. We focus on a model where the agent weights originate from a stochastic process, resulting in weight uncertainty. We analyze the expected effect of the quota on voting power given the weight generating process. We examine two extreme cases of the balls and bins model: uniform and exponentially decaying probabilities. We show that the choice of a quota may have a large influence on the power disparity of the agents, even when the governing distribution is likely to result in highly similar weights for the agents. We characterize various interesting repetitive fluctuation patterns in agents' power as a function of the quota.
【Keywords】:
【Paper Link】 【Pages】:81-87
【Authors】: Yoram Bachrach ; Omer Lev ; Yoad Lewenberg ; Yair Zick
【Abstract】: Voting systems in which voters are partitioned to districts encourage accountability by providing voters an easily identifiable district representative, but can result in a selection of representatives not representative of the electorate's preferences. In some cases, a party may have a majority of the popular vote, but lose the elections due to districting effects. We define the Misrepresentation Ratio which quantifies the deviation from proportional representation in a district-based election, and provide bounds for this ratio under various voting rules. We also examine probabilistic models for election outcomes, and provide an algorithm for approximating the expected Misrepresentation Ratio under a given probabilistic election model. Finally, we provide simulation results for several such probabilistic election models, showing the effects of the number of voters and candidates on the misrepresentation ratio.
【Keywords】:
【Paper Link】 【Pages】:88-94
【Authors】: Nathanaël Barrot ; Jérôme Lang
【Abstract】: Several methods exist for making collective decisions on a set of variables when voters possibly have preferential dependencies. None is based on approval voting. We define a family of rules for approval-based voting on combinatorial domains, where voters cast conditional approval ballots, allowing them to approve values of a variable conditionally on the values of other variables. We study three such rules. The first two generalize simple multiwinner approval voting and minimax approval voting. The third one is an approval-based version of sequential voting on combinatorial domains. We study some properties of these rules, and compare their outcomes.
【Keywords】:
【Paper Link】 【Pages】:95-101
【Authors】: Francesco Belardinelli ; Andreas Herzig
【Abstract】: Recently logics for strategic ability have gained pre-eminence in the modelisation and analysis of game-theoretic scenarios. In this paper we provide a contribution to the comparison of two popular frameworks: Concurrent Game Structures (CGS) and Coalition Logic of Propositional Control (CLPC). Specifically, we ground the abstract abilities of agents in CGS on Propositional Control, thus obtaining a class of CGS that has the same expressive power as CL-PC. We study the computational properties of this setting. Further, we relax some of the assumptions of CL-PC so as to introduce a wider class of computationally-grounded CGS.
【Keywords】:
【Paper Link】 【Pages】:102-108
【Authors】: Bernhard Bliem ; Robert Bredereck ; Rolf Niedermeier
【Abstract】: We study the problem of finding a Pareto-efficient and envy-free allocation of a set of indivisible resources to a set of agents with monotonic preferences, either dichotomous or additive. Motivated by results of Bouveret and Lang [JAIR 2008], we provide a refined computational complexity analysis by studying the influence of three natural parameters: the number n of agents, the number m of resources, and the number z of different numbers occurring in utility-based preferences of the agents. On the negative side, we show that small values for n and z alone do not significantly lower the computational complexity in most cases. On the positive side, devising fixed-parameter algorithms we show that all considered problems are tractable in case of small m. Furthermore, we develop a fixed-parameter algorithm indicating that the problem with additive preferences becomes computationally tractable in case of small n and small z.
【Keywords】:
【Paper Link】 【Pages】:109-115
【Authors】: Édouard Bonnet
【Abstract】: Durak is a Russian card game in which players try to get rid of all their cards via a particular attack/defense mechanism. The last player standing with cards loses. We show that, even restricted to the perfect information two-player game, finding optimal moves is a hard problem. More precisely, we prove that, given a generalized Durak position, it is PSPACE-complete to decide if a player has a winning strategy. We also show that deciding if an attack can be answered is NP-hard.
【Keywords】:
【Paper Link】 【Pages】:116-122
【Authors】: Florian Brandl ; Felix Brandt ; Christian Geist
【Abstract】: Two important requirements when aggregating the preferences of multiple agents are that the outcome should be economically efficient and the aggregation mechanism should not be manipulable. In this paper, we provide a computer-aided proof of a sweeping impossibility using these two conditions for randomized aggregation mechanisms. More precisely, we show that every efficient aggregation mechanism can be manipulated for all expected utility representations of the agents' preferences. This settles a conjecture by Aziz et al. [2013b] and strengthens a number of existing theorems, including statements that were shown within the special domain of assignment. Our proof is obtained by formulating the claim as a satisfiability problem over predicates from real-valued arithmetic, which is then checked using an SMT (satisfiability modulo theories) solver. To the best of our knowledge, this is the first application of SMT solvers in computational social choice.
【Keywords】:
【Paper Link】 【Pages】:123-129
【Authors】: Simina Brânzei ; Yuezhou Lv ; Ruta Mehta
【Abstract】: Single minded agents have strict preferences, in which a bundle is acceptable only if it meets a certain demand. Such preferences arise naturally in scenarios such as allocating computational resources among users, where the goal is to fairly serve as many requests as possible. In this paper we study the fair division problem for such agents, which is complex due to discontinuity and complementarities of preferences. Our solution concept — the competitive allocation from equal incomes (CAEI) — is inspired from market equilibria and implements fair outcomes through a pricing mechanism. We study existence and computation of CAEI for multiple divisible goods, discrete goods, and cake cutting. Our solution is useful more generally, when the players have a target set of goods, and very small positive values for any bundle other than their target set.
【Keywords】:
【Paper Link】 【Pages】:130-136
【Authors】: Markus Brill ; Edith Elkind ; Ulle Endriss ; Umberto Grandi
【Abstract】: We introduce a model of preference diffusion in which agents in a social network update their preferences based on those of their influencers in the network, and we study the dynamics of this model. Preferences are modelled as ordinal rankings over a finite set of alternatives. At each time step, some of the agents update the relative ordering of two alternatives adjacent in their current ranking with the majority view of their influencers. We consider both a synchronous and an asynchronous variant of this model. Our results show how the graph-theoretic structure of the social network and the structure of the agents' preferences affect the termination of the diffusion process and the properties of the preference profile at the time of termination.
【Keywords】:
【Paper Link】 【Pages】:137-143
【Authors】: Qingpeng Cai ; Aris Filos-Ratsikas ; Pingzhong Tang
【Abstract】: We study the problem of locating a public facility on a real line or an interval, when agents' costs are their (expected) distances from the location of the facility. Our goal is to minimize the maximum envy over all agents, which we will refer to as the minimax envy objective, while at the same time ensuring that agents will report their most preferred locations truthfully. First, for the problem of locating the facility on a real line, we propose a class of truthful-in-expectation mechanisms that generalize the well-known LRM mechanism, the best of which has performance arbitrarily close to the social optimum. Then, we restrict the possible locations of the facility to a real interval and consider two cases; when the interval is determined relatively to the agents' reports and when the interval is fixed in advance. For the former case, we prove that for any choice of such an interval, there is a mechanism in the aforementioned class with additive approximation arbitrarily close to the best approximation achieved by any truthful-in-expectation mechanism. For the latter case, we prove that the approximation of the best truthful-in-expectation mechanism is between 1/3 and 1/2.
【Keywords】:
【Paper Link】 【Pages】:144-150
【Authors】: Ioannis Caragiannis ; Laurent Gourvès ; Jérôme Monnot
【Abstract】: We study the optimization problem of designing the program of a conference with parallel sessions, so that the intended participants are as happy as possible from the talks they can attend. Interestingly, this can be thought of as a two-dimensional extension of a scheme proposed by Chamberlin and Courant [1983] for achieving proportional representation in multi-winner elections. We show that different variations of the problem are computationally hard by exploiting relations of the problem with well-known hard graph problems. On the positive side, we present polynomial-time algorithms that compute conference programs that have a social utility that is provably close to the optimal one (within constant factors). Our algorithms are either combinatorial or based on linear programming and randomized rounding.
【Keywords】:
【Paper Link】 【Pages】:151-157
【Authors】: Ioannis Caragiannis ; Swaprava Nath ; Ariel D. Procaccia ; Nisarg Shah
【Abstract】: How should one aggregate ordinal preferences expressed by voters into a measurably superior social choice? A well-established approach — which we refer to as implicit utilitarian voting — assumes that voters have latent utility functions that induce the reported rankings, and seeks voting rules that approximately maximize utilitarian social welfare. We extend this approach to the design of rules that select a subset of alternatives. We derive analytical bounds on the performance of optimal (deterministic as well as randomized) rules in terms of two measures, distortion and regret. Empirical results show that regret-based rules are more compelling than distortion-based rules, leading us to focus on developing a scalable implementation for the optimal (deterministic) regret-based rule. Our methods underlie the design and implementation of an upcoming social choice website.
【Keywords】:
【Paper Link】 【Pages】:158-164
【Authors】: Mithun Chakraborty ; Sanmay Das
【Abstract】: Prediction markets are popular mechanisms for aggregating information about a future event. In situations where market participants may significantly influence the outcome, running the prediction market could change the incentives of participants in the process that creates the outcome. We propose a new game-theoretic model that captures two aspects of real-world prediction markets: (1) agents directly affect the outcome the market is predicting, (2) some outcome-deciders may not participate in the market. We show that this game has two types of equilibria: When some outcome-deciders are unlikely to participate in the market, equilibrium prices reveal expected market outcomes conditional on participants' private information, whereas when all outcome-deciders are likely to participate, equilibria are collusive — agents effectively coordinate in an uninformative and untruthful way.
【Keywords】:
【Paper Link】 【Pages】:165-171
【Authors】: Hau Chan ; Albert Xin Jiang
【Abstract】: Congestion games are a well-studied class of games that has been used to model real-world systems such as Internet routing. In many congestion games, each player's number of strategies can be exponential in the natural description of the game. Most existing algorithms for game theoretic computation, from computing expected utilities and best responses to finding Nash equilibrium and other solution concepts, all involve enumeration of pure strategies. As a result, such algorithms would take exponential time on these congestion games. In this work, we study congestion games in which each player's strategy space can be described compactly using a set of linear constraints. For instance, network congestion games naturally fall into this subclass as each player's strategy can be described by a set of flow constraints. We show that we can represent any mixed strategy compactly using marginals which specify the probability of using each resource. As a consequence, the expected utilities and the best responses can be computed in polynomial time. We reduce the problem of computing a best/worst symmetric approximate mixed-strategy Nash equilibrium in symmetric congestion games to a constraint optimization problem on a graph formed by the resources and the strategy constraints. As a result, we present a fully polynomial time approximation scheme (FPTAS) for this problem when the graph has bounded tree width.
【Keywords】:
【Paper Link】 【Pages】:172-179
【Authors】: Krishnendu Chatterjee ; Rasmus Ibsen-Jensen ; Josef Tkadlec
【Abstract】: Balanced knockout tournaments are ubiquitous in sports competitions and are also used in decision-making and elections. The traditional computational question, that asks to compute a draw (optimal draw) that maximizes the winning probability for a distinguished player, has received a lot of attention. Previous works consider the problem where the pairwise winning probabilities are known precisely, while we study how robust is the winning probability with respect to small errors in the pairwise winning probabilities. First, we present several illuminating examples to establish: (a) there exist deterministic tournaments (where the pairwise winning probabilities are 0 or 1) where one optimal draw is much more robust than the other; and (b) in general, there exist tournaments with slightly suboptimal draws that are more robust than all the optimal draws. The above examples motivate the study of the computational problem of robust draws that guarantee a specified winning probability. Second, we present a polynomial-time algorithm for approximating the robustness of a draw for sufficiently small errors in pairwise winning probabilities, and obtain that the stated computational problem is NP-complete. We also show that two natural cases of deterministic tournaments where the optimal draw could be computed in polynomial time also admit polynomial-time algorithms to compute robust optimal draws.
【Keywords】:
【Paper Link】 【Pages】:180-186
【Authors】: Taolue Chen ; Fu Song ; Zhilin Wu
【Abstract】: In this paper, we investigate model checking algorithms for variants of strategy logic over pushdown multi-agent systems, modeled by pushdown game structures (PGSs). We consider various fragments of strategy logic, i.e., SL[CG], SL[DG], SL[1G] and BSIL. We show that the model checking problems on PGSs for SL[CG], SL[DG] and SL[1G] are 3EXTIME-complete, which are not harder than the problem for the subsumed logic ATL*. When BSIL is concerned, the model checking problem becomes 2EXPTIME-complete. Our algorithms are automata-theoretic and based on the saturation technique, which are amenable to implementations.
【Keywords】:
【Paper Link】 【Pages】:187-193
【Authors】: Yukun Cheng ; Xiaotie Deng ; Qi Qi ; Xiang Yan
【Abstract】: In this paper, we consider the popular proportional sharing mechanism and discuss the incentives and opportunities of an agent to lie for personal gains in resource exchange game. The main result is a proof that an agent manipulating the proportional sharing mechanism by misreporting its resource amount will not benefit its own utility eventually. This result establishes a strategic stability property of the resource exchange protocol. We further illustrate and confirm the result via network examples.
【Keywords】:
【Paper Link】 【Pages】:194-200
【Authors】: Yun Kuen Cheung
【Abstract】: We revisit the problem of designing strategyproof mechanisms for allocating divisible items among two agents who have linear utilities, where payments are disallowed and there is no prior information on the agents' preferences. The objective is to design strategyproof mechanisms which are competitive against the most efficient (but not strategyproof) mechanism. For the case with two items: We provide a set of sufficient conditions for strategyproofness. We use an analytic approach to derive strategyproof mechanisms which are more competitive than all prior strategyproof mechanisms. We improve the linear-program-based proof of Guo and Conitzer to show new upper bounds on competitive ratios. We provide the first compact proof on upper bound of competitiveness. For the cases with any number of items, we build on the Partial Allocation mechanisms introduced by Cole et al. to design a strategyproof mechanism which is 0.67776-competitive, breaking the 2/3 barrier. We also propose a new sub-class of strategyproof mechanisms for any numbers of agents and items, which we call it Dynamic-Increasing-Price mechanisms, where each agent purchases the items using virtual money, and the prices of the items depend on other agents' preferences.
【Keywords】:
【Paper Link】 【Pages】:201-207
【Authors】: Ross Conroy ; Yifeng Zeng ; Jing Tang
【Abstract】: Interactive dynamic influence diagrams~(I-DIDs) provide an explicit way of modeling how a subject agent solves decision making problems in the presence of other agents in a common setting. To optimize its decisions, the subject agent needs to predict the other agents' behavior, that is generally obtained by solving their candidate models. This becomes extremely difficult since the model space may be rather large, and grows when the other agents act and observe over the time. A recent proposal for solving I-DIDs lies in a concept of value equivalence (VE) that shows potential advances on significantly reducing the model space. In this paper, we establish a principled framework to implement the VE techniques and propose an approximate method to compute VE of candidate models. The development offers ample opportunity of exploiting VE to further improve the scalability of I-DID solutions. We theoretically analyze properties of the approximate techniques and show empirical results in multiple problem domains.
【Keywords】:
【Paper Link】 【Pages】:208-214
【Authors】: Natalia Criado ; Jose M. Such
【Abstract】: Real-world norm monitors have limited capabilities for observing agents. This paper proposes a novel mechanism to take full advantage of limited observation capabilities by selecting the agents to be monitored. Our evaluation shows this significantly increases the number of violations detected.
【Keywords】:
【Paper Link】 【Pages】:215-221
【Authors】: Palash Dey ; Neeldhara Misra
【Abstract】: Eliciting preferences of a set of agents over a set of items is a problem of fundamental interest in artificial intelligence in general and social choice theory in particular. Prior works on preference elicitation focus on unrestricted domain and the domain of single peaked preferences and show that the preferences in single peaked domain can be elicited by much less number of queries compared to unrestricted domain. We extend this line of research and study preference elicitation for single peaked preferences on trees which is a strict superset of the domain of single peaked preferences. We show that the query complexity crucially depends on the number of leaves, the path cover number, and the distance from path of the underlying single peaked tree, whereas the other natural parameters like maximum degree, diameter, pathwidth do not play any direct role in determining query complexity. We then investigate the query complexity for finding a weak Condorcet winner for preferences single peaked on a tree and show that this task has much less query complexity than preference elicitation. Here again we observe that the number of leaves in the underlying single peaked tree and the path cover number of the tree influence the query complexity of the problem.
【Keywords】:
【Paper Link】 【Pages】:222-228
【Authors】: Palash Dey ; Neeldhara Misra
【Abstract】: Eliciting the preferences of a set of agents over a set of alternatives is a problem of fundamental importance in social choice theory. Prior work on this problem has studied the query complexity of preference elicitation for the unrestricted domain and for the domain of single peaked preferences. In this paper, we consider the domain of single crossing preference profiles and study the query complexity of preference elicitation under various settings. We consider two distinct situations: when an ordering of the voters with respect to which the profile is single crossing is known versus when it is unknown. We also consider random and sequential access models. The main contribution of our work is to provide polynomial time algorithms with low query complexity for preference elicitation in all the above cases. Further, we show that the query complexities of our algorithms are optimal up to constant factors for all but one of the above cases.
【Keywords】:
【Paper Link】 【Pages】:229-235
【Authors】: Palash Dey ; Neeldhara Misra ; Y. Narahari
【Abstract】: The Coalitional Manipulation problem has been studied extensively in the literature for many voting rules. However, most studies have focused on the complete information setting, wherein the manipulators know the votes of the non-manipulators. While this assumption is reasonable for purposes of showing intractability, it is unrealistic for algorithmic considerations. In most real-world scenarios, it is impractical for the manipulators to have accurate knowledge of all the other votes. In this paper, we investigate manipulation with incomplete information. In our framework, the manipulators know a partial order for each voter that is consistent with the true preference of that voter. In this setting, we formulate three natural computational notions of manipulation, namely weak, opportunistic, and strong manipulation. We consider several scenarios for which the traditional manipulation problems are easy (for instance, Borda with a single manipulator). For many of them, the corresponding manipulative questions that we propose turn out to be computationally intractable. Our hardness results often hold even when very little information is missing, or in other words, even when the instances are quite close to the complete information setting. Our overall conclusion is that computational hardness continues to be a valid obstruction to manipulation, in the context of a more realistic model.
【Keywords】:
【Paper Link】 【Pages】:236-242
【Authors】: Ulle Endriss ; Svetlana Obraztsova ; Maria Polukarov ; Jeffrey S. Rosenschein
【Abstract】: Classical results in social choice theory on the susceptibility of voting rules to strategic manipulation make the assumption that the manipulator has complete information regarding the preferences of the other voters. In reality, however, voters only have incomplete information, which limits their ability to manipulate. We explore how these limitations affect both the manipulability of voting rules and the dynamics of systems in which voters may repeatedly update their own vote in reaction to the moves made by others. We focus on the Plurality, Veto, k-approval, Borda, Copeland, and Maximin voting rules, and consider several types of information that are natural in the context of these rules, namely information on the current front-runner, on the scores obtained by each alternative, and on the majority graph induced by the individual preferences.
【Keywords】:
【Paper Link】 【Pages】:243-249
【Authors】: Piotr Faliszewski ; Arkadii M. Slinko ; Nimrod Talmon
【Abstract】: We study a combinatorial problem formulated in terms of the following group-formation scenario. Given some agents, where each agent has preferences over the set of potential group leaders, the task is to partition the agents into groups and assign a group leader to each of them, so that the group leaders have as high support as possible from the groups they are assigned to lead. We model this scenario as a voting problem, where the goal is to partition a set of voters into a prescribed number of groups so that each group elects its leader, i.e., their leader is a unique winner in the corresponding election. We study the computational complexity of this problem (and several of its variants) for Approval elections.
【Keywords】:
【Paper Link】 【Pages】:250-256
【Authors】: Piotr Faliszewski ; Piotr Skowron ; Arkadii Slinko ; Nimrod Talmon
【Abstract】: We consider several natural classes of committee scoring rules, namely, weakly separable, representation-focused, top-k-counting, OWA-based, and decomposable rules. We study some of their axiomatic properties, especially properties of monotonicity, and concentrate on containment relations between them. We characterize SNTV, Bloc, and k-approval Chamberlin-Courant, as the only rules in certain intersections of these classes. We introduce decomposable rules, describe some of their applications, and show that the class of decomposable rules strictly contains the class of OWA-based rules.
【Keywords】:
【Paper Link】 【Pages】:257-263
【Authors】: Piotr Faliszewski ; Laurent Gourvès ; Jérôme Lang ; Julien Lesca ; Jérôme Monnot
【Abstract】: We consider a Plurality-voting scenario, where the candidates are split between parties, and each party nominates exactly one candidate for the final election. We study the computational complexity of deciding if there is a set of nominees such that a candidate from a given party wins in the final election. In our second problem, the goal is to decide if a candidate from a given party always wins, irrespective who is nominated. We show that these problems are computationally hard, but are polynomial-time solvable for restricted settings.
【Keywords】:
【Paper Link】 【Pages】:264-270
【Authors】: Wenyi Fang ; Pingzhong Tang ; Song Zuo
【Abstract】: Over the past decade, computer-automated barter exchange has become one of the most successful applications at the intersection of AI and economics. Standard exchange models, such as house allocation and kidney exchange cannot be applied to an emerging industrial application, coined digital good exchange, where an agent still possesses her initial endowment after exchanging with others. However, her valuation toward her endowment decreases as it is possessed by more agents. We put forward game theoretical models tailored for digital good exchange. In the first part of the paper, we first consider a natural class of games where agents can choose either a subset of other participants' items or no participation at all. It turns out that this class of games can be modeled as a variant of congestion games. We prove that it is in general NP-complete to determine whether there exists a non-trivial pure Nash equilibrium where at least some agent chooses a nonempty subset of items. However, we show that in a subset of games for single-minded agents with unit demand, there exist non-trivial Pure Nash equilibria and put forward an efficient algorithm to find such equilibria. In the second part of the paper, we investigate digital good exchange from a mechanism design perspective. We ask if there is a truthful mechanism in this setting that can achieve good social welfare guarantee. To this end, we design a randomized fixed-price-exchange mechanism that is individually rational and truthful, and for two-player case yields a tight log-approximation with respect to any individually rational allocation.
【Keywords】:
【Paper Link】 【Pages】:271-278
【Authors】: Paolo Felli ; Brian Logan ; Sebastian Sardiña
【Abstract】: A key problem in the manufacture of highly-customized products is the synthesis of controllers able to manufacture any instance of a given product type on a given production or assembly line. In this paper, we extend classical AI behavior composition to manufacturing settings. We first introduce a novel solution concept for manufacturing composition, target production processes, that are able to manufacture multiple instances of a product simultaneously in a given production plant. We then propose a technique for synthesizing the largest target production process, together with an associated controller for the machines in the plant.
【Keywords】:
【Paper Link】 【Pages】:279-285
【Authors】: Dimitris Fotakis ; Dimitris Palyvos-Giannas ; Stratis Skoulakis
【Abstract】: We study convergence properties of opinion dynamics with local interactions and limited information exchange. We adopt a general model where the agents update their opinions in rounds to a weighted average of the opinions in their neighborhoods. For fixed neighborhoods, we present a simple randomized protocol that converges in expectation to the stable state of the Friedkin-Johnsen model. For opinion-dependent neighborhoods, we show that the Hegselmann-Krause model converges to a stable state if each agent's neighborhood is restricted either to a subset of her acquaintances or to a small random subset of agents. Our experimental findings indicate that for a wide range of parameters, the convergence time and the number of opinion clusters of the neighborhood-restricted variants are comparable to those of the standard Hegselmann-Krause model.
【Keywords】:
【Paper Link】 【Pages】:286-293
【Authors】: Enrico H. Gerding ; Sebastian Stein ; Sofia Ceppi ; Valentin Robu
【Abstract】: Vehicle-to-grid (V2G) is a promising approach whereby electric vehicles (EVs) are used to store excess electricity supply (e.g., from renewable sources), which is sold back to the grid in times of scarcity. In this paper we consider the setting of a smart car park, where EVs come and go, and can be used for V2G while parked. We develop novel allocation and payment mechanisms which truthfully elicit the EV owners' preferences and constraints, including arrival, departure, required charge, as well as the costs of discharging due to loss of efficiency of the battery. The car park will schedule the charging and discharging of each EV, ensuring the constraints of the EVs are met, and taking into consideration predictions about future electricity prices. Optimally solving the global problem is intractable, and we present three novel heuristic online scheduling algorithms. We show that, under certain conditions, two of these satisfy monotonicity and are therefore truthful. We furthermore evaluate the algorithms using simulations, and we show that some of our algorithms benefit significantly from V2G, achieving positive benefit for the car park even when agents do not pay for using it.
【Keywords】:
【Paper Link】 【Pages】:294-300
【Authors】: Julio Godoy ; Ioannis Karamouzas ; Stephen J. Guy ; Maria L. Gini
【Abstract】: Multi-agent navigation methods typically assume that all agents use the same underlying framework to navigate to their goal while avoiding colliding with each other. However, such assumption does not hold when agents do not know how other agents will move.We address this issue by proposing a Bayesian inference approach where an agent estimates the navigation model and goal of each neighbor, and uses this to compute a plan that minimizes collisions while driving it to its goal. Simulation experiments performed in many scenarios demonstrate that an agent using our approach computes safer and more time-efficient paths as compared to those generated without our inference approach anda state-of-the-art local navigation framework.
【Keywords】:
【Paper Link】 【Pages】:301-307
【Authors】: Umberto Grandi ; Paolo Turrini
【Abstract】: We study a rating system in which a set of individuals (e.g., the customers of a restaurant) evaluate a given service (e.g, the restaurant), with their aggregated opinion determining the probability of all individuals to use the service and thus its generated revenue. We explicitly model the influence relation by a social network, with individuals being influenced by the evaluation of their trusted peers. On top of that we allow a malicious service provider (e.g., the restaurant owner) to bribe some individuals, i.e., to invest a part of his or her expected income to modify their opinion, therefore influencing his or her final gain. We analyse the effect of bribing strategies under various constraints, and we show under what conditions the system is bribery-proof, i.e., no bribing strategy yields a strictly positive expected gain to the service provider.
【Keywords】:
【Paper Link】 【Pages】:308-314
【Authors】: Nika Haghtalab ; Fei Fang ; Thanh Hong Nguyen ; Arunesh Sinha ; Ariel D. Procaccia ; Milind Tambe
【Abstract】: State-of-the-art applications of Stackelberg security games — including wildlife protection — offer a wealth of data, which can be used to learn the behavior of the adversary. But existing approaches either make strong assumptions about the structure of the data, or gather new data through online algorithms that are likely to play severely suboptimal strategies. We develop a new approach to learning the parameters of the behavioral model of a bounded rational attacker (thereby pinpointing a near optimal strategy), by observing how the attacker responds to only three defender strategies. We also validate our approach using experiments on real and synthetic data.
【Keywords】:
【Paper Link】 【Pages】:315-321
【Authors】: Xiaowei Huang ; Qingliang Chen ; Jie Meng ; Kaile Su
【Abstract】: Reactive agents are suitable for representing physical resources in manufacturing control systems. An important challenge of agent-based manufacturing control systems is to develop formal and structured approaches to support their specification and verification. This paper proposes a logic-based approach, by generalising that of model checking multiagent systems, for the reconfigurability of reactive multiagent systems. Two reconfigurability scenarios are studied, for the resulting system being a monolithic system or an individual module, and their computational complexity results are given.
【Keywords】:
【Paper Link】 【Pages】:322-328
【Authors】: Hooyeon Lee ; Ashish Goel
【Abstract】: Consider an event organizer who is trying to schedule a group meeting. Availability of agents is unknown to the organizer a priori, but the organizer may have probability estimates on availability of each agent for each date/time option. The organizer can ask an agent to reveal her availability, but it causes inconvenience for the agent, and thus the organizer wishes to find an agreeable outcome at a minimum number of such queries. Motivated by this example, we study the Probabilistic Matrix Inspection problem in which we are given a matrix of Bernoulli random variables that are mutually independent, and the objective is to determine whether the matrix contains a column consisting only of 1's. We are allowed to inspect an arbitrary entry at unit cost, which reveals the realization of the entry, and we wish to find an inspection policy whose expected number of inspections is minimum. We first show that an intuitive greedy algorithm exists for 1-row and 1-column matrices, and we generalize this to design an algorithm that finds an optimal policy in polynomial time for the general case.
【Keywords】:
【Paper Link】 【Pages】:329-337
【Authors】: Yuqian Li ; Vincent Conitzer ; Dmytro Korzhyk
【Abstract】: Algorithms for computing game-theoretic solutions have recently been applied to a number of security domains. However, many of the techniques developed for compact representations of security games do not extend to Bayesian security games, which allow us to model uncertainty about the attacker's type. In this paper, we introduce a general framework of catcher-evader games that can capture Bayesian security games as well as other game families of interest. We show that computing Stackelberg strategies is NP-hard, but give an algorithm for computing a Nash equilibrium that performs well in experiments. We also prove that the Nash equilibria of these games satisfy the interchangeability property, so that equilibrium selection is not an issue.
【Keywords】:
【Paper Link】 【Pages】:338-344
【Authors】: Thanasis Lianeas ; Evdokia Nikolova ; Nicolás E. Stier Moses
【Abstract】: We consider a nonatomic selfish routing model with independent stochastic travel times for each edge, represented by mean and variance latency functions that depend on arc flows. This model can apply to traffic in the Internet or in a road network. Variability negatively impacts packets or drivers, by introducing jitter in transmission delays which lowers quality of streaming audio or video, or by making it more difficult to predict the arrival time at destination. The price of risk aversion (PRA) has been defined as the worst-case ratio of the cost of an equilibrium with risk-averse players who seek risk-minimizing paths, and that of an equilibrium with risk-neutral users who minimize the mean travel time of a path [Nikolova and Stier-Moses, 2015]. This inefficiency metric captures the degradation of system performance caused by variability and risk aversion. In this paper, we provide the first lower bounds on the PRA. First, we show a family of structural lower bounds, which grow linearly with the size of the graph and players' risk-aversion. They are tight for graph sizes that are powers of two. We also provide asymptotically tight functional bounds that depend on the allowed latency functions but not on the topology. The functional bounds match the price-of-anarchy bounds for congestion games multiplied by an extra factor that accounts for risk aversion. Finally, we provide a closed-form formula of the PRA for a family of graphs that generalize series-parallel graphs and the Braess graph. This formula also applies to the mean-standard deviation user objective a much more complex model of risk-aversion due to the cost of a path being non-additive over edge costs.
【Keywords】:
【Paper Link】 【Pages】:345-351
【Authors】: Hongyao Ma ; Reshef Meir ; David C. Parkes
【Abstract】: The existence of truthful social choice mechanisms strongly depends on whether monetary transfers are allowed. Without payments there are no truthful, non-dictatorial mechanisms under mild requirements, whereas the VCG mechanism guarantees truthfulness along with welfare maximization when there are payments and utility is quasi-linear in money. In this paper we study mechanisms in which we can use payments but where agents have non quasi-linear utility functions. Our main result extends the Gibbard-Satterthwaite impossibility result by showing that, for two agents, the only truthful mechanism for at least three alternatives under general decreasing utilities remains dictatorial. We then show how to extend the VCG mechanism to work under a more general utility space than quasi-linear (the "parallel domain) and show that the parallel domain is maximal—no mechanism with the VCG properties exists in any larger domain.
【Keywords】:
【Paper Link】 【Pages】:352-358
【Authors】: Hongyao Ma ; Valentin Robu ; Na Li ; David C. Parkes
【Abstract】: We study the problem of incentivizing reliable demand-response in modern electricity grids. Each agent is uncertain about her future ability to reduce demand and unreliable. Agents who choose to participate in a demand-response scheme may be paid when they respond and penalized otherwise. The goal is to reliably achieve a demand reduction target while selecting a minimal set of agents from those willing to participate. We design incentive-aligned, direct and indirect mechanisms. The direct mechanism elicits both response probabilities and costs, while the indirect mechanism elicits willingness to accept a penalty in the case of non-response. We benchmark against a spot auction, in which demand reduction is purchased from agents when needed. Both the direct and indirect mechanisms achieve the reliability target in a dominant-strategy equilibrium, select a small number of agents to prepare, and do so at low cost and with much lower variance in payments than the spot auction.
【Keywords】:
【Paper Link】 【Pages】:359-365
【Authors】: Erika Mackin ; Lirong Xia
【Abstract】: We initiate a research agenda of mechanism design for categorized domain allocation problems (CDAPs), where indivisible items from multiple categories are allocated to agents without monetary transfer and each agent gets at least one item per category. We focus on basic CDAPs, where each agent gets exactly one item per category. We first characterize serial dictatorships by a minimal set of three axiomatic properties: strategy-proofness, non-bossiness, and category-wise neutrality. Then, we propose a natural extension of serial dictatorships called categorial sequential allocation mechanisms (CSAMs), which allocate the items in multiple rounds: in each round, the designated agent chooses an item from a designated category. We fully characterize the worst-case rank efficiency of CSAMs for optimistic and pessimistic agents.
【Keywords】:
【Paper Link】 【Pages】:366-372
【Authors】: Debmalya Mandal ; David C. Parkes
【Abstract】: We study the social choice problem where a group of n voters report their preferences over alternatives and a voting rule is used to select an alternative. We show that when the preferences of voters are positively correlated according to the Kendall-Tau distance, the probability that any scoring rule is not ex post incentive compatible (EPIC) goes to zero exponentially fast with the number of voters, improving over the previously known rate of 1/√n for independent preferences. Motivated by rank-order models from machine learning, we introduce two examples of positively-correlated models, namely Conditional Mallows and Conditional Plackett-Luce. Conditional Mallows satisfies Kendall-Tau correlation and fits our positive result. We also prove that Conditional Plackett-Luce becomes EPIC exponentially quickly.
【Keywords】:
【Paper Link】 【Pages】:373-379
【Authors】: Mehdi Mashayekhi ; Hongying Du ; George F. List ; Munindar P. Singh
【Abstract】: In a multiagent system, a (social) norm describes what the agents may expect from each other. Norms promote autonomy (an agent need not comply with a norm) and heterogeneity (a norm describes interactions at a high level independent of implementation details). Researchers have studied norm emergence through social learning where the agents interact repeatedly in a graph structure. In contrast, we consider norm emergence in an open system, where membership can change, and where no predetermined graph structure exists. We propose Silk, a mechanism wherein a generator monitors interactions among member agents and recommends norms to help resolve conflicts. Each member decides on whether to accept or reject a recommended norm. Upon exiting the system, a member passes its experience along to incoming members of the same type. Thus, members develop norms in a hybrid manner to resolve conflicts. We evaluate Silk via simulation in the traffic domain. Our results show that social norms promoting conflict resolution emerge in both moderate and selfish societies via our hybrid mechanism.
【Keywords】:
【Paper Link】 【Pages】:380-386
【Authors】: Georgios Methenitis ; Michael Kaisers ; Han La Poutré
【Abstract】: Current electricity tariffs for retail rarely provide incentives for intelligent demand response of flexible customers. Such customers could otherwise contribute to balancing supply and demand in future smart grids. This paper proposes an innovative risk-sharing tariff to incentivize intelligent customer behavior. A two-step parameterized payment scheme is proposed, consisting of a prepayment based on the expected consumption, and a supplementary payment for any observed deviation from the anticipated consumption. Within a game-theoretical analysis, we capture the strategic conflict of interest between a retailer and a customer in a two-player game, and we present optimal, i.e., best response, strategies for both players in this game. We show analytically that the proposed tariff provides customers of varying flexibility with variable incentives to assume and alleviate a fraction of the balancing risk, contributing in this way to the uncertainty reduction in the envisioned smart-grid.
【Keywords】:
【Paper Link】 【Pages】:387-393
【Authors】: Vahab S. Mirrokni ; Renato Paes Leme ; Pingzhong Tang ; Song Zuo
【Abstract】: Consider a buyer with independent additive valuations for a set of goods, and a seller who is constrained to sell one item at a time in an online fashion. If the seller is constrained to run independent auctions for each item, then he would run Myerson's optimal auction for each item. If the seller is allowed to use the full power of dynamic mechanism design and have the auction for each item depend on the outcome of the previous auctions, he is able to perform much better. The main issues in implementing such strategies in online settings where items arrive over time are that the auction might be too complicated or it makes too strong assumptions on the buyer's rationality or seller's commitment over time. This motivates us to explore a restricted family of dynamic auctions that can be implemented in an online fashion and without too much commitment from the seller ahead of time. In particular, we study a set of auction in which the space of single-shot auctions is augmented with a structure that we call bank account, a real number for each node that summarizes the history so far. This structure allows the seller to store deficits or surpluses of buyer utility from each individual auction and even them out on the long run. This is akin to enforcing individual rationality constraint on average rather than per auction. We also study the effect of enforcing a maximum limit to the values that bank account might grow, which means that we enforce that besides the auction being individually rational on average it is also not far from being individually rational at any given interval. Interestingly, even with these restrictions, we can achieve significantly better revenue and social welfare compared to separate Myerson auctions.
【Keywords】:
【Paper Link】 【Pages】:394-400
【Authors】: Reuth Mirsky ; Ya'akov (Kobi) Gal
【Abstract】: Plan Recognition algorithms require to recognize a complete hierarchy explaining the agent's actions and goals. While the output of such algorithms is informative to the recognizer, the cost of its calculation is high in run-time, space, and completeness. Moreover, performing plan recognition online requires the observing agent to reason about future actions that have not yet been seen and maintain a set of hypotheses to support all possible options. This paper presents a new and efficient algorithm for online plan recognition called SLIM (Semi-Lazy Inference Mechanism). It combines both a bottom-up and top-down parsing processes, which allow it to commit only to the minimum necessary actions in real-time, but still provide complete hypotheses post factum. We show both theoretically and empirically that although the computational cost of this process is still exponential, there is a significant improvement in run-time when compared to a state of the art of plan recognition algorithm.
【Keywords】:
【Paper Link】 【Pages】:401-407
【Authors】: Reuth Mirsky ; Roni Stern ; Ya'akov (Kobi) Gal ; Meir Kalech
【Abstract】: Plan recognition algorithms infer agents' plans from their observed actions. Due to imperfect knowledge about the agent's behavior and the environment, it is often the case that there are multiple hypotheses about an agent's plans that are consistent with the observations, though only one of these hypotheses is correct. This paper addresses the problem of how to disambiguate between hypotheses, by querying the acting agent about whether a candidate plan in one of the hypotheses matches its intentions. This process is performed sequentially and used to update the set of possible hypotheses during the recognition process. The paper defines the sequential plan recognition process (SPRP), which seeks to reduce the number of hypotheses using a minimal number of queries. We propose a number of policies for the SPRP which use maximum likelihood and information gain to choose which plan to query. We show this approach works well in practice on two domains from the literature, significantly reducing the number of hypotheses using fewer queries than a baseline approach. Our results can inform the design of future plan recognition systems that interleave the recognition process with intelligent interventions of their users.
【Keywords】:
【Paper Link】 【Pages】:408-415
【Authors】: Jayanth Krishna Mogali ; Stephen F. Smith ; Zachary B. Rubinstein
【Abstract】: We propose a new distributed algorithm for decoupling the Multiagent Simple Temporal Network (MaSTN) problem. The agents cooperatively decouple the MaSTN while simultaneously optimizing a sum of concave objectives local to each agent. Several schedule flexibility measures are applicable in this framework. We pose the MaSTN decoupling problem as a distributed convex optimization problem subject to constraints having a block angular structure; we adapt existing variants of Alternating Direction Method of Multiplier (ADMM) type methods to perform decoupling optimally. The resulting algorithm is an iterative procedure that is guaranteed to converge. Communication only takes place between agents with temporal inter-dependences and the information exchanged between them is carried out in a privacy preserving manner. We present experimental results for the proposed method on problems of varying sizes, and demonstrate its effectiveness in terms of solving quality and computational cost.
【Keywords】:
【Paper Link】 【Pages】:416-423
【Authors】: Catherine Moon ; Vincent Conitzer
【Abstract】: In multiagent systems, often agents need to be assigned to different roles. Multiple aspects should be taken into account for this, such as agents' skills and constraints posed by existing assignments. In this paper, we focus on another aspect: when the agents are self-interested, careful role assignment is necessary to make cooperative behavior an equilibrium of the repeated game. We formalize this problem and provide an easy-to-check necessary and sufficient condition for a given role assignment to induce cooperation. However, we show that finding whether such a role assignment exists is in general NP-hard. Nevertheless, we give two algorithms for solving the problem. The first is based on a mixed-integer linear program formulation. The second is based on a dynamic program, and runs in pseudopolynomial time if the number of agents is constant. Minor modifications of these algorithms also allow for determination of the minimal subsidy necessary to induce cooperation. In our experiments, the IP performs much, much faster.
【Keywords】:
【Paper Link】 【Pages】:424-432
【Authors】: Dmitry Moor ; Sven Seuken ; Tobias Grubenmann ; Abraham Bernstein
【Abstract】: In some auction domains, there is uncertainty regarding the final availability of the goods being auctioned off. For example, a government may auction off spectrum from its public safety network, but it may need this spectrum back in times of emergency. In such a domain, standard combinatorial auctions perform poorly because they lead to violations of individual rationality (IR), even in expectation, and to very low efficiency. In this paper, we study the design of core-selecting payment rules for such domains. Surprisingly, we show that in this new domain, there does not exist a payment rule with is guaranteed to be ex-post core-selecting. However, we show that by designing rules that are execution-contingent, i.e., by charging payments that are conditioned on the realization of the availability of the goods, we can reduce IR violations. We design two core-selecting rules that always satisfy IR in expectation. To study the performance of our rules we perform a computational Bayes-Nash equilibrium analysis. We show that, in equilibrium, our new rules have better incentives, higher efficiency, and a lower rate of ex-post IR violations than standard core-selecting rules.
【Keywords】:
【Paper Link】 【Pages】:433-439
【Authors】: Harikrishna Narasimhan ; Shivani Agarwal ; David C. Parkes
【Abstract】: We use statistical machine learning to develop methods for automatically designing mechanisms in domains without money. Our goal is to find a mechanism that best approximates a given target function subject to a design constraint such as strategy-proofness or stability. The proposed approach involves identifying a rich parametrized class of mechanisms that resemble discriminant-based multiclass classifiers, and relaxing the resulting search problem into an SVM-style surrogate optimization problem. We use this methodology to design strategy-proof mechanisms for social choice problems with single-peaked preferences, and stable mechanisms for two-sided matching problems. To the best of our knowledge, ours is the first automated approach for designing stable matching rules. Experiments on synthetic and real-world data confirm the usefulness of our methods.
【Keywords】:
【Paper Link】 【Pages】:440-446
【Authors】: Svetlana Obraztsova ; Zinovi Rabinovich ; Edith Elkind ; Maria Polukarov ; Nicholas R. Jennings
【Abstract】: Trembling hand (TH) equilibria were introduced by Selten in 1975. Intuitively, these are Nash equilibria that remain stable when players assume that there is a small probability that other players will choose off-equilibrium strategies. This concept is useful for equilibrium refinement, i.e., selecting the most plausible Nash equilibria when the set of all Nash equilibria can be very large, as is the case, for instance, for Plurality voting with strategic voters. In this paper, we analyze TH equilibria of Plurality voting. We provide an efficient algorithm for computing a TH best response and establish many useful properties of TH equilibria in Plurality voting games. On the negative side, we provide an example of a Plurality voting game with no TH equilibria, and show that it is NP-hard to check whether a given Plurality voting game admits a TH equilibrium where a specific candidate is among the election winners.
【Keywords】:
【Paper Link】 【Pages】:447-453
【Authors】: Steven Okamoto ; Roie Zivan ; Aviv Nahon
【Abstract】: The Distributed Breakout Algorithm (DBA) is a local search algorithm that was originally designed to solve DisCSPs and DisMaxCSPs. Extending it to general-valued DCOPs requires three design choices: manner of modifying base costs (multiplicative weights or additive penalties); definition of constraint violation (non-zero cost, non-minimum cost, and maximum cost); and scope of modifying cost tables during breakout (entry, row, column, or table). We propose Generalized DBA (GDBA) to span the 24 combinations in the three dimensions. In our theoretical analysis we prove that some variants of GDBA are equivalent for certain problems, and prove that other variants may find suboptimal solutions even on tree topologies where DBA is complete. Our extensive empirical evaluation on various benchmarks shows that in practice, GDBA is capable of finding solutions of equal or significantly lower cost than alternative heuristic approaches (including DSA).
【Keywords】:
【Paper Link】 【Pages】:454-460
【Authors】: Tenda Okimoto ; Tony Ribeiro ; Damien Bouchabou ; Katsumi Inoue
【Abstract】: Team formation is the problem of selecting a group of agents, where each agent has a set of skills; the aim is to accomplish a given mission (a set of tasks), where each task is made precise by a skill necessary for managing it. In a dynamic environment that offers the possibility of losing agents during a mission, e.g., some agents break down, the robustness of a team is crucial. In this paper, the focus is laid on the mission oriented robust multi-team formation problem. A formal framework is defined and two algorithms are provided to tackle this problem, namely, a complete and an approximate algorithm. In the experiments, these two algorithms are evaluated in RMASBench (a rescue multi-agent benchmarking platform used in the RoboCup Rescue Simulation League).We empirically show that (i) the approximate algorithm is more realistic for RMASBench compared to the complete algorithm and (ii) considering the robust mission multi-teams have a better control on the fire spread than the sophisticate solvers provided in RMASBench.
【Keywords】:
【Paper Link】 【Pages】:461-467
【Authors】: James Parker ; Maria L. Gini
【Abstract】: We propose solutions for assignment of physical tasks to heterogeneous agents when the costs of the tasks change over time. We assume tasks have a natural growth rate which is counteracted by the work applied by the agents. As the future cost of a task depends on the agents allocation, reasoning must be both spatial and temporal to effectively minimize the growth so tasks can be completed. We present optimal solutions for two general classes of growth functions and heuristic solutions for other cases. Empirical results are given in RoboCup Rescue for agents with different capabilities.
【Keywords】:
【Paper Link】 【Pages】:468-474
【Authors】: Pierre Rust ; Gauthier Picard ; Fano Ramparany
【Abstract】: We consider environments in which smart devices equipped with limited communication and computation capabilities have to cooperate to self-configure their state in an energy-efficient manner, as to meet user-defined requirements. Such requirements are expressed as scene rules, configured by the user using an intuitive interface that connects conditions on sensors' and actuators' states and actions on actuators. We translate this smart environment configuration problem into a constraint optimization problem. As to install distributiveness, robustness, and openness, we solve it using distributed message-passing algorithms. We illustrate our approach through a running example, and evaluate the performances of the implemented protocols on a simulated realistic environment.
【Keywords】:
【Paper Link】 【Pages】:475-481
【Authors】: Wen Shen ; Cristina V. Lopes ; Jacob W. Crandall
【Abstract】: With proper management, Autonomous Mobility-on-Demand (AMoD) Systems have great potential to satisfy the transport demand of urban populations by providing safe, convenient, and affordable ridesharing services. Meanwhile, such systems can substantially decrease private car ownership and use, and thus significantly reduce traffic congestion, energy consumption, and carbon emissions. To achieve this objective, an AMoD system requires private information about the demand from passengers. However, due to self-interestedness, passengers are unlikely to cooperate with the service providers in this regard. Therefore, an online mechanism is desirable if it incentivizes passengers to truthfully report their actual demand. For the purpose of promoting ridesharing, we hereby introduce a posted-price, integrated online ridesharing mechanism (IORS) that satisfies desirable properties such as ex-post incentive compatibility, individual rationality, and budget-balance. Numerical results indicate the competitiveness of IORS compared with two benchmarks, namely the optimal assignment and an offline, auction-based mechanism.
【Keywords】:
【Paper Link】 【Pages】:482-488
【Authors】: Sunil Simon ; Dominik Wojtczak
【Abstract】: We study strategic games on weighted directed graphs, where the payoff of a player is defined as the sum of the weights on the edges from players who chose the same strategy augmented by a fixed non-negative bonus for picking a given strategy. These games capture the idea of coordination in the absence of globally common strategies. Prior work shows that the problem of determining the existence of a pure Nash equilibrium for these games is NP-complete already for graphs with all weights equal to one and no bonuses. However, for several classes of graphs (e.g. DAGs and cliques) pure Nash equilibria or even strong equilibria always exist and can be found by simply following a particular improvement or coalition-improvement path, respectively. In this paper we identify several natural classes of graphs for which a finite improvement or coalition-improvement path of polynomial length always exists, and, as a consequence, a Nash equilibrium or strong equilibrium in them can be found in polynomial time. We also argue that these results are optimal in the sense that in natural generalisations of these classes of graphs, a pure Nash equilibrium may not even exist.
【Keywords】:
【Paper Link】 【Pages】:489-495
【Authors】: Warut Suksompong
【Abstract】: We consider an assignment problem that has aspects of fair division as well as social choice. In particular, we investigate the problem of assigning a small subset from a set of indivisible items to multiple players so that the chosen subset is agreeable to all players, i.e., every player weakly prefers the chosen subset to any subset of its complement. For an arbitrary number of players, we derive tight upper bounds on the size for which a subset of that size that is agreeable to all players always exists when preferences are monotonic. We then present polynomial-time algorithms that find an agreeable subset of approximately half of the items when there are two or three players and preferences are responsive. Our results translate to a 2-approximation on the individual welfare of every player when preferences are subadditive.
【Keywords】:
【Paper Link】 【Pages】:496-502
【Authors】: Tamir Tassa ; Roie Zivan ; Tal Grinshpoun
【Abstract】: Region-optimal algorithms are local search algorithms for the solution of Distributed ConstraintOptimization Problems (DCOPs). In each iteration of the search in such algorithms, every agent selects a group of agents that comply with some selection criteria (each algorithm specifies different criteria). Then, the agent who selected the group, called the mediator, collects assignment information from the group and neighboring agents outside the group, in order to find an optimal set of assignments for its group's agents. A contest between mediators of adjacent groups determines which groups will replace their assignments in that iteration to the found optimal ones. In this work we present a framework called RODA (Region-Optimal DCOP Algorithm) that encompasses the algorithms in the region-optimality family, and in particular any method for selecting groups. We devise a secure implementation of RODA, called PRODA, which preserves constraint privacy and partial decision privacy. The two main cryptographic means that enable this privacy preservation are secret sharing and homomorphic encryption. We estimate the computational overhead of P-RODA with respect to RODA and give an upper bound that depends on the group and domain sizes and the graph topology but not on the number of agents. The estimations are backed with experimental results.
【Keywords】:
【Paper Link】 【Pages】:503-509
【Authors】: Nicolas Troquard
【Abstract】: We introduce a class of resource games where resources and preferences are described with the language of a resource-sensitive logic. We present two decision problems, the first of which is deciding whether an action profile is a Nash equilibrium. When dealing with resources, interesting questions arise as to whether some undesirable equilibria can be eliminated by a central authority by redistributing the available resources among the agents. We will thus study the decision problem of rational elimination. We will consider them in the contexts of dichotomous or pseudo-dichotomous preferences, and of logics that admit or not the weakening rule. This will offer a variety of complexity results that are applicable to a large number of settings.
【Keywords】:
【Paper Link】 【Pages】:510-516
【Authors】: Elaine Wah ; Sébastien Lahaie ; David M. Pennock
【Abstract】: In this paper, we employ simulation-based methods to study the role of a market maker in improving price discovery in a prediction market. In our model, traders receive a lagged signal of a ground truth, which is based on real price data from prediction markets on NBA games in the 2014-2015 season. We employ empirical game-theoretic analysis to identify equilibria under different settings of market maker liquidity and spread. We study two settings: one in which traders only enter the market once, and one in which traders have the option to reenter to trade later. We evaluate welfare and the profits accrued by traders, and we characterize the conditions under which the market maker promotes price discovery in both settings.
【Keywords】:
【Paper Link】 【Pages】:517-523
【Authors】: Kyle Hollins Wray ; Luis Enrique Pineda ; Shlomo Zilberstein
【Abstract】: Semi-Autonomous Systems (SAS) encapsulate a stochastic decision process explicitly controlled by both an agent and a human, in order to leverage the distinct capabilities of each actor. Planning in SAS must address the challenge of transferring control quickly, safely, and smoothly back-and-forth between the agent and the human. We formally define SAS and the requirements to guarantee that the controlling entities are always able to act competently. We then consider applying the model to Semi-Autonomous VEhicles (SAVE), using a hierarchical approach in which micro-level transfer-of-control actions are governed by a high-fidelity POMDP model. Macro-level path planning in our hierarchical approach is performed by solving a Stochastic Shortest Path (SSP) problem. We analyze the integrated model and show that it provides the required guarantees. Finally, we test the SAVE model using real-world road data from Open Street Map (OSM) within 10 cities, showing the benefits of the collaboration between the agent and human.
【Keywords】:
【Paper Link】 【Pages】:524-530
【Authors】: Feng Wu ; Sarvapali D. Ramchurn ; Xiaoping Chen
【Abstract】: We consider a disaster response scenario where emergency responders have to complete rescue tasks in dynamic and uncertain environment with the assistance of multiple UAVs to collect information about the disaster space. To capture the uncertainty and partial observability of the domain, we model this problem as a POMDP. However, the resulting model is computationally intractable and cannot be solved by most existing POMDP solvers due to the large state and action spaces. By exploiting the problem structure we propose a novel online planning algorithm to solve this model. Specifically, we generate plans for the responders based on Monte-Carlo simulations and compute actions for the UAVs according to the value of information. Our empirical results confirm that our algorithm significantly outperforms the state-of-the-art both in time and solution quality.
【Keywords】:
【Paper Link】 【Pages】:531-537
【Abstract】: Coral reefs are valuable and fragile ecosystems which are under threat from human activities like coral mining. Many countries have built marine protected areas (MPAs) and protect their ecosystems through boat patrol. However, it remains a significant challenge to efficiently patrol the MPAs given the limited patrol resources of the protection agency and potential destructors' strategic actions. In this paper, we view the problem of efficiently patrolling for protecting coral reef ecosystems from a game-theoretic perspective and propose 1) a new Stackelberg game model to formulate the problem of protecting MPAs, 2) two algorithms to compute the efficient protection agency's strategies: CLP in which the protection agency's strategies are compactly represented as fractional flows in a network, and CDOG which combines the techniques of compactly representing defender strategies and incrementally generating strategies. Experimental results show that our approach leads to significantly better solution quality than that of previous works.
【Keywords】:
【Paper Link】 【Pages】:538-545
【Authors】: Yue Yin ; Yevgeniy Vorobeychik ; Bo An ; Noam Hazon
【Abstract】: Election control encompasses attempts from an external agent to alter the structure of an election in order to change its outcome. This problem is both a fundamental theoretical problem in social choice, and a major practical concern for democratic institutions.Consequently, this issue has received considerable attention, particularly as it pertains to different voting rules. In contrast, the problem of how election control can be prevented or deterred has been largely ignored. We introduce the problem of optimal protection against election control, where manipulation is allowed at the granularity of groups of voters (e.g., voting locations), through a denial-of-service attack, and the defender allocates limited protection resources to prevent control. We show that for plurality voting, election control through group deletion to prevent a candidate from winning is in P, while it is NP-Hard to prevent such control. We then present a double-oracle framework for computing an optimal prevention strategy, developing exact mixed-integer linear programming for mulations for both the defender and attacker oracles (both of these subproblems we show to be NP-Hard), as well as heuristic oracles. Experiments conducted on both synthetic and real data demonstrate that the proposed computational framework can scale to realistic problem instances.
【Keywords】:
【Paper Link】 【Pages】:546-553
【Authors】: Ruohan Zhang ; Yue Yu ; Mahmoud El Chamie ; Behçet Açikmese ; Dana H. Ballard
【Abstract】: This paper studies a decision-making problem for heterogeneous multi-agent systems with safety density constraints. An individual agent's decision-making problem is modeled by the standard Markov Decision Process (MDP) formulation. However, an important special case occurs when the MDP states may have limited capacities, hence upper bounds on the expected number of agents in each state are imposed. We refer to these upper bound constraints as "safety" constraints. If agents follow unconstrained policies (policies that do not impose the safety constraints), the safety constraints might be violated. In this paper, we devise algorithms that provide safe decision-making policies. The set of safe decision policies can be shown to be convex, and hence the policy synthesis is tractable via reliable and fast Interior Point Method (IPM) algorithms. We evaluate the effectiveness of proposed algorithms first using a simple MDP, and then using a dynamic traffic assignment problem. The numerical results demonstrate that safe decision-making algorithms in this paper significantly outperform other baselines.
【Keywords】:
【Paper Link】 【Pages】:554-560
【Authors】: Steven Adriaensen ; Ann Nowé
【Abstract】: To date, algorithms for real-world problems are most commonly designed following a manual, ad-hoc, trial and error approach, making algorithm design a tedious, time-consuming and costly process. Recently, Programming by Optimization (PbO) has been proposed as an alternative design paradigm in which algorithmic choices are left open by design and algorithm configuration methods (e.g. ParamILS) are used to automatically generate the best algorithm for a specific use-case. We argue that, while powerful, contemporary configurators limit themselves by abstracting information that can otherwise be exploited to speed up the optimization process as well as improve the quality of the resulting design. In this work, we propose an alternative white box approach, reformulating the algorithm design problem as a Markov Decision Process, capturing the intrinsic relationships between design decisions and their respective contribution to overall algorithm performance. Subsequently, we discuss and illustrate the benefits of this formulation experimentally.
【Keywords】:
【Paper Link】 【Pages】:561-567
【Authors】: Zaheen Farraz Ahmad ; Robert C. Holte ; Michael Bowling
【Abstract】: Curling is an adversarial two-player game with a continuous state and action space, and stochastic transitions. This paper focuses on one aspect of the full game, namely, finding the optimal "hammer shot," which is the last action taken before a score is tallied. We survey existing methods for finding an optimal action in a continuous, low-dimensional space with stochastic outcomes, and adapt a method based on Delaunay Triangulation to our application. Experiments using our curling physics simulator show that the adapted Delaunay Triangulation's shot selection outperforms other algorithms, and with some caveats, exceeds Olympic-level human performance.
【Keywords】:
【Paper Link】 【Pages】:568-574
【Authors】: Shaowei Cai ; Jinkun Lin
【Abstract】: This paper explores techniques for fast solving the maximum weight clique problem (MWCP) in very large scale real-world graphs. Because of the size of such graphs and the intractability of MWCP, previously developed algorithms may not be applicable. Although recent heuristic algorithms make progress in solving MWCP in massive graphs, they still need considerable time to get a good solution. In this work, we propose a new method for MWCP which interleaves between clique construction and graph reduction. We also propose three novel ideas to make it efficient, and develop an algorithm called FastWClq. Experiments on massive graphs from various applications show that, FastWClq finds better solutions than state of the art algorithms while the run time is much less. Further, FastWClq proves the optimal solution for about half of the graphs in an averaged time less than one second.
【Keywords】:
【Paper Link】 【Pages】:575-581
【Authors】: Thomas Guyet ; Yves Moinard ; Jacques Nicolas ; René Quiniou
【Abstract】: This paper describes an application of Answer Set Programming (ASP) to crop allocation for generating realistic landscapes. The aim is to cover optimally a bare landscape, represented by its plot graph, with spatial patterns describing local arrangements of crops. This problem belongs to the hard class of graph packing problems and is modeled in the framework of ASP. The approach provides a compact solution to the basic problem and at the same time allows extensions such as a flexible integration of expert knowledge. Particular attention is paid to the treatment of symmetries, especially due to sub-graph isomorphism issues. Experiments were conducted on a database of simulated and real landscapes. Currently, the approach can process graphs of medium size, a size that enables studies on real agricultural practices.
【Keywords】:
【Paper Link】 【Pages】:582-588
【Authors】: David Hofmeyr
【Abstract】: Genetic algorithms are stochastic search heuristics which are popular for their broad applicability, especially in combinatorial search problems. The search mechanism relies on an abstraction of genetic evolution and selection as seen in nature. This paper introduces a topological structure for the search space which is consistent with existing theory and practice for genetic algorithms, namely forma analysis. A notion of convexity is defined within this context and connections between this definition and forma analysis are established. This framework provides an alternative perspective on the exploitation/exploration dilemma as well as population convergence, which relates directly to the genetic operators employed to drive the evolution process. It also provides a different interpretation of design constraints associated with genetic algorithm implementations. The intention is to provide a new analytical perspective for genetic algorithms, and to establish a connection with exact search methods through the concept of convexity. Genetic algorithms are stochastic search heuristics which are popular for their broad applicability, especially in combinatorial search problems. The search mechanism relies on an abstraction of genetic evolution and selection as seen in nature. This paper introduces a topological structure for the search space which is consistent with existing theory and practice for genetic algorithms, namely forma analysis. A notion of convexity is defined within this context and connections between this definition and forma analysis are established. This framework provides an alternative perspective on the exploitation/exploration dilemma as well as population convergence, which relates directly to the genetic operators employed to drive the evolution process. It also provides a different interpretation of design constraints associated with genetic algorithm implementations. The intention is to provide a new analytical perspective for genetic algorithms, and to establish a connection with exact search methods through the concept of convexity.
【Keywords】:
【Paper Link】 【Pages】:589-595
【Authors】: Takashi Imamichi ; Takayuki Osogami ; Rudy Raymond
【Abstract】: We study the problem of map-matching, or finding the route on a road network from a trace of noisy and sparse observed points, particularly when a huge number of points are given. The algorithms based on Hidden Markov Models (HMMs) are known to achieve high accuracy for noisy and sparse data but suffer from high computational cost. We find that the bottleneck of the HMM-based map-matching is in the shortest path search for calculating transition probabilities. We propose a technique to truncate the shortest path search before finding all the shortest paths in the HMM-based map-matching without losing accuracy. We run the one-to-many shortest path searches on the reversed network and terminate the searches based on the log likelihood of the Viterbi algorithm. Computational experiments show that the proposed approaches can reduce the computational cost by a factor of at least 5.4.
【Keywords】:
【Paper Link】 【Pages】:596-602
【Authors】: Joachim Jansen ; Bart Bogaerts ; Jo Devriendt ; Gerda Janssens ; Marc Denecker
【Abstract】: Inductive definitions and justifications are well-studied concepts. Solvers that support inductive definitions have been developed, but several of their computationally nice properties have never been exploited to improve these solvers. In this paper, we present a new notion called relevance. We determine a class of literals that are relevant for a given definition and partial interpretation, and show that choices on irrelevant atoms can never benefit the search for a model. We propose an early stopping criterion and a modification of existing heuristics that exploit relevance. We present a first implementation in MinisatID and experimentally evaluate our approach, and study how often existing solvers make choices on irrelevant atoms.
【Keywords】:
【Paper Link】 【Pages】:603-609
【Authors】: Kustaa Kangas ; Teemu Hankala ; Teppo Mikael Niinimäki ; Mikko Koivisto
【Abstract】: We present two algorithms for computing the number of linear extensions of a given n-element poset. Our first approach builds upon an O(2n n)-time dynamic programming algorithm by splitting subproblems into connected components and recursing on them independently. The recursion may run over two alternative subproblem spaces, and we provide heuristics for choosing the more efficient one. Our second algorithm is based on variable elimination via inclusion-exclusion and runs in time O(nt+4)), where t is the treewidth of the cover graph. We demonstrate experimentally that these new algorithms outperform previously suggested ones for a wide range of posets, in particular when the posets are sparse.
【Keywords】:
【Paper Link】 【Pages】:610-616
【Authors】: Richard E. Korf
【Abstract】: We compare sorting and hashing for implicit graph search using disk storage. We first describe efficient pipelined implementations of both algorithms, which reduce disk I/O. We then compare the two algorithms and find that hashing is faster, but that sorting requires less disk storage. We also compare disk-based with in-memory search, and surprisingly find that there is little or no time overhead associated with disk-based search. We present experimental results on the sliding-tile puzzles, Rubik's Cube, and the 4-peg Towers of Hanoi.
【Keywords】:
【Paper Link】 【Pages】:617-623
【Authors】: Javier Larrosa ; Emma Rollon ; Rina Dechter
【Abstract】: Many combinatorial problems are solved with a Depth-First search (DFS) guided by a heuristic and it is well-known that this method is very fragile with respect to heuristic mistakes. One standard way to make DFS more robust is to search by increasing number of discrepancies. This approach has been found useful in several domains where the search structure is a height-bounded OR tree. In this paper we investigate the generalization of discrepancy-based search to AND/OR search trees and propose an extension of the Limited Discrepancy Search (LDS) algorithm. We demonstrate the relevance of our proposal in the context of Graphical Models. In these problems, which can be solved with either a standard OR search tree or an AND/OR tree, we show the superiority of our approach. For a fixed number of discrepancies, the search space visited by the AND/OR algorithm strictly contains the search space visited by standard LDS, and many more nodes can be visited due to the multiplicative effect of the AND/OR decomposition. Besides, if the AND/OR tree achieves a significant size reduction with respect to the standard OR tree, the cost of each iteration of the AND/OR algorithm is asymptotically lower than in standard LDS. We report experiments on the minsum problem on different domains and show that the AND/OR version of LDS usually obtains better solutions given the same CPU time.
【Keywords】:
【Paper Link】 【Pages】:624-630
【Authors】: Kangwei Liu ; Junge Zhang ; Peipei Yang ; Kaiqi Huang
【Abstract】: Recently, MRFs with two-dimensional (2D) labels have proved useful to many applications, such as image matching and optical flow estimation. Due to the huge 2D label set in these problems, existing optimization algorithms tend to be slow for the inference of 2D label MRFs, and this greatly limits the practical use of 2D label MRFs. To solve the problem, this paper presents an efficient algorithm, named FastLCD. Unlike previous popular move-making algorithms (e.g., α-expansion) that visit all the labels exhaustively in each step, FastLCD optimizes the 2D label MRFs by performing label coordinate descents alternately in horizontal, vertical and diagonal directions, and by this way, it does not need to visit all the labels exhaustively. FastLCD greatly reduces the search space of the label set and benefits from a lower time complexity. Experimental results show that FastLCD is much faster, while it still yields high quality results.
【Keywords】:
【Paper Link】 【Pages】:631-638
【Authors】: Ciaran McCreesh ; Patrick Prosser ; James Trimble
【Abstract】: We show how to generate "really hard'" random instances for subgraph isomorphism problems. For the non-induced variant, we predict and observe a phase transition between satisfiable and unsatisfiable instances, with a corresponding complexity peak seen in three different solvers. For the induced variant, much richer behaviour is observed, and constrainedness gives a better measure of difficulty than does proximity to a phase transition. We also discuss variable and value ordering heuristics, and their relationship to the expected number of solutions.
【Keywords】:
【Paper Link】 【Pages】:639-646
【Authors】: Ole J. Mengshoel ; Youssef Ahres ; Tong Yu
【Abstract】: Stochastic local search (SLS) algorithms have proven to be very competitive in solving hard computational problems.This paper investigates the foundations of SLS algorithms. We develop a simple SLS algorithm, MarkovSLS, with three search operators: greedy, noise, and restart. The search operators are controlled by probability parameters, leading to soft (probabilistic) rather than hard (deterministic) restarts. We consider two special cases of the MarkovSLS algorithm: SoftSLS and AdaptiveSLS. In SoftSLS, the probability parameters are fixed, enabling analysis using standard homogeneous Markov chains. We study the interaction between the restart and noise parameters in SoftSLS, and optimize them analytically in addition to the traditional empirical approach. Experimentally, we investigate the dependency of SoftSLS's performance on its noise and restart parameters, validating the analytical results. AdaptiveSLS dynamically adjusts its noise and restart parameters during search. Experimentally, on synthetic and feature selection problems, we compare AdaptiveSLS with other algorithms including an analytically optimized version of SoftSLS, and find that it performs well while not requiring prior knowledge of the search space.
【Keywords】:
【Paper Link】 【Pages】:647-654
【Authors】: Abdelkader Ouali ; Samir Loudni ; Yahia Lebbah ; Patrice Boizumault ; Albrecht Zimmermann ; Lakhdar Loukil
【Abstract】: Conceptual clustering combines two long-standing machine learning tasks: the unsupervised grouping of similar instances and their description by symbolic concepts. In this paper, we decouple the problems of finding descriptions and forming clusters by first mining formal concepts (i.e. closed itemsets), and searching for the best k clusters that can be described with those itemsets. Most existing approaches performing the two steps separately are of a heuristic nature and produce results of varying quality. Instead, we address the problem of finding an optimal constrained conceptual clustering by using integer linear programming techniques. Most other generic approaches for this problem tend to have problems scaling. Our approach takes advantageous of both techniques, the general framework of integer linear programming, and high-speed specialized approaches of data mining. Experiments performed on UCI datasets show that our approach efficiently finds clusterings of consistently high quality.
【Keywords】:
【Paper Link】 【Pages】:655-661
【Authors】: Fei Peng ; Tuomas Sandholm
【Abstract】: As publishers gather more information about their users, they can use that information to enable advertisers to create increasingly targeted campaigns. This enables better usage of advertising inventory. However, it also dramatically increases the complexity that the publisher faces when optimizing campaign admission decisions and inventory allocation to campaigns.We develop an optimal anytime algorithm for abstracting fine-grained audience segments into coarser abstract segments that are not too numerous for use in such optimization. Compared to the segment abstraction algorithm by Walsh et al. [2010] for the same problem, it yields two orders of magnitude improvement in run time and significant improvement in abstraction quality. These benefits hold both for guaranteed and non-guaranteed campaigns. The performance stems from three improvements: 1) a quadratic-time (as opposed to doubly exponential or heuristic) algorithm for finding an optimal split of an abstract segment, 2) a better scoring function for evaluating splits, and 3) splitting time lossily like any other targeting attribute (instead of losslessly segmenting time first).
【Keywords】:
【Paper Link】 【Pages】:662-668
【Authors】: André Grahl Pereira ; Robert Holte ; Jonathan Schaeffer ; Luciana S. Buriol ; Marcus Ritt
【Abstract】: We present a novel admissible pattern database heuristic (D) and tie-breaking rule (L) for Sokoban, allowing us to increase the number of optimally solved standard Sokoban instances from 20 to 28 and the number of proved optimal solutions from 25 to 32 compared to previous methods. The previously best heuristic for Sokoban (I) used the idea of an intermediate goal state to enable the effective use of pattern database heuristics in transportation domains, where the mapping of movable objects to goal locations is not fixed beforehand. We extend this idea to allow the use of multiple intermediate goal states and show that heuristic I is no longer effective. We solve this problem and show that our heuristic D is effective in this situation. Sokoban is a well-known single-agent search domain characterized by a large branching factor, long solution lengths, and the presence of unsolvable states. Given the exponential growth in the complexity of standard Sokoban instances, the increase in the number of optimally solved instances represents a major advance in our understanding of how to search in extremely large search spaces.
【Keywords】:
【Paper Link】 【Pages】:669-675
【Authors】: Masoud Safilian ; S. Mehdi Hashemi ; Sepehr Eghbali ; Aliakbar Safilian
【Abstract】: The subpath planning problem (SPP) is a branch of path planning problem, which has widespread applications in automated manufacturing process as well as vehicle and robot navigation. This problem aims to find the shortest path or tour subject to covering a set of given subpaths. By casting SPP to a graph routing problem, we propose a deterministic 2-approximation algorithm finding near optimal solutions, which runs in O(n3) time.
【Keywords】:
【Paper Link】 【Pages】:676-682
【Authors】: Nathan R. Sturtevant ; Jingwei Chen
【Abstract】: This paper studies external memory bidirectional search. That is, how bidirectional search algorithms can run using external memory such as hard drives or solid state drives. While external memory algorithms have been broadly studied in unidirectional search, they have not been studied in the context of bidirectional search. We show that the primary bottleneck in bidirectional search is the question of solution detection — knowing when the two search frontiers have met. We propose a method of delayed solution detection that makes external bidirectional search more efficient. Experimental results show the effectiveness of the approach.
【Keywords】:
【Paper Link】 【Pages】:683-689
【Authors】: Nathan R. Sturtevant ; Steve Rabin
【Abstract】: Jump Point Search, an algorithm developed for fast search on uniform cost grids, has successfully improved the performance of grid-based search. But, the approach itself is actually a set of diverse ideas applied together. This paper decomposes the algorithm and gradually reconstructs it, showing the component pieces from which the algorithm is constructed. In this process, we are able to define a spectrum of new algorithms that borrow and repurpose ideas from Jump Point Search. This decomposition opens the door for applying the ideas from Jump Point Search in other grid domains with significantly different characteristics from two dimensional grids.
【Keywords】:
【Paper Link】 【Pages】:690-697
【Authors】: Timothy Yee ; Viliam Lisý ; Michael H. Bowling
【Abstract】: Real world applications of artificial intelligence often require agents to sequentially choose actions from continuous action spaces with execution uncertainty. When good actions are sparse, domain knowledge is often used to identify a discrete set of promising actions. These actions and their uncertain effects are typically evaluated using a recursive search procedure. The reduction of the problem to a discrete search problem causes severe limitations, notably, not exploiting all of the sampled outcomes when evaluating actions, and not using outcomes to help find new actions outside the original set. We propose a new Monte Carlo tree search (MCTS) algorithm specifically designed for exploiting an execution model in this setting. Using kernel regression, it generalizes the information about action quality between actions and to unexplored parts of the action space. In a high fidelity simulator of the olympic sport of curling, we show that this approach significantly outperforms existing MCTS methods.
【Keywords】:
【Paper Link】 【Pages】:698-704
【Authors】: Robin Arcangioli ; Christian Bessiere ; Nadjib Lazaar
【Abstract】: QUACQ is a constraint acquisition system that assists a non-expert user to model her problem as a constraint network by classifying (partial) examples as positive or negative. For each negative example, QUACQ focuses onto a constraint of the target network. The drawback is that the user may need to answer a great number of such examples to learn all the constraints. In this paper, we provide a new approach that is able to learn a maximum number of constraints violated by a given negative example. Finally we give an experimental evaluation that shows that our approach improves on QUACQ.
【Keywords】:
【Paper Link】 【Pages】:705-711
【Authors】: Christian Bessiere ; Emmanuel Hebrard ; George Katsirelos ; Zeynep Kiziltan ; Toby Walsh
【Abstract】: We need to reason about rankings of objects in a wide variety of domains including information retrieval, sports tournaments, bibliometrics, and statistics. We propose a global constraint therefore for modeling rankings. One important application for rankings is in reasoning about the correlation or uncorrelation between two sequences. For example, we might wish to have consecutive delivery schedules correlated to make it easier for clients and employees, or uncorrelated to avoid predictability and complacence. We therefore also consider global correlation constraints between rankings. For both ranking and correlation constraints, we propose efficient filtering algorithms and decompositions, and report experimental results demonstrating the promise of our proposed approach.
【Keywords】:
【Paper Link】 【Pages】:712-719
【Authors】: Chris Cameron ; Holger H. Hoos ; Kevin Leyton-Brown
【Abstract】: A Virtual Best Solver (VBS) is a hypothetical algorithm that selects the best solver from a given portfolio of alternatives on a per-instance basis. The VBS idealizes performance when all solvers in a portfolio are run in parallel, and also gives a valuable bound on the performance of portfolio-based algorithm selectors. Typically, VBS performance is measured by running every solver in a portfolio once on a given instance and reporting the best performance over all solvers. Here, we argue that doing so results in a flawed measure that is biased to reporting better performance when a randomized solver is present in an algorithm portfolio. As long as there is more than 1 solver in the portfolio, a single randomized solver can cause VBS bias. Specifically, this flawed notion of VBS tends to show performance better than that achievable by a perfect selector that for each given instance runs the solver with the best expected running time. We report results from an empirical study using solvers and instances submitted to several SAT competitions, in which we observe significant bias on many random instances and some combinatorial instances. We also show that the bias increases with the number of randomized solvers and decreases as we average solver performance over many independent runs per instance. We propose an alternative VBS performance measure by (1) empirically obtaining the solver with best expected performance for each instance and (2) taking bootstrap samples for this solver on every instance, to obtain a confidence interval on VBS performance. Our findings shed new light on widely studied algorithm selection benchmarks and help explain performance gaps observed between VBS and state-of-the-art algorithm selection approaches.
【Keywords】:
【Paper Link】 【Pages】:720-726
【Authors】: Abderrazak Daoudi ; Younes Mechqrane ; Christian Bessiere ; Nadjib Lazaar ; El-Houssine Bouyakhf
【Abstract】: Constraint acquisition systems assist the non-expert user in modeling her problem as a constraint network. Most existing constraint acquisition systems interact with the user by asking her to classify an example as positive or negative. Such queries do not use the structure of the problem and can thus lead the user to answer a large number of queries. In this paper, we propose Predict&Ask, an algorithm based on the prediction of missing constraints in the partial network learned so far. Such missing constraints are directly asked to the user through recommendation queries, a new, more informative kind of queries. Predict&Ask can be plugged in any constraint acquisition system. We experimentally compare the QuAcq system to an extended version boosted by the use of our recommendation queries. The results show that the extended version improves the basic QuAcq.
【Keywords】:
【Paper Link】 【Pages】:727-734
【Authors】: Jeffrey M. Dudek ; Kuldeep S. Meel ; Moshe Y. Vardi
【Abstract】: The runtime performance of modern SAT solvers on random k-CNF formulas is deeply connected with the 'phase-transition' phenomenon seen empirically in the satisfiability of random k-CNF formulas. Recent universal hashing-based approaches to sampling and counting crucially depend on the runtime performance of SAT solvers on formulas expressed as the conjunction of both k-CNF and XOR constraints (known as k-CNF-XOR formulas), but the behavior of random k-CNF-XOR formulas is unexplored in prior work. In this paper, we present the first study of the satisfiability of random k-CNF-XOR formulas. We show empirical evidence of a surprising phase-transition that follows a linear trade-off between k-CNF and XOR constraints. Furthermore, we prove that a phase-transition for k-CNF-XOR formulas exists for k = 2 and (when the number of k-CNF constraints is small) for k > 2.
【Keywords】:
【Paper Link】 【Pages】:735-743
【Authors】: Azadeh Farzan ; Zachary Kincaid
【Abstract】: Satisfiability-checking of formulas in the theory of linear rational arithmetic (LRA) has broad applications including program verification and synthesis. Satisfiability Modulo Theories (SMT) solvers are effective at checking satisfiability of the ground fragment of LRA, but applying them to quantified formulas requires a costly quantifier elimination step. This article presents a novel decision procedure for LRA that leverages SMT solvers for the ground fragment of LRA, but avoids explicit quantifier elimination. The intuition behind the algorithm stems from an interpretation of a quantified formula as a game between two players, whose goals are to prove that the formula is either satisfiable or not. The algorithm synthesizes a winning strategy for one of the players by iteratively improving candidate strategies for both. Experimental results demonstrate that the proposed procedure is competitive with existing solvers.
【Keywords】:
【Paper Link】 【Pages】:744-750
【Authors】: Zeynep Kiziltan ; Marco Lippi ; Paolo Torroni
【Abstract】: Modeling in constraint programming is a hard task that requires considerable expertise. Automated model reformulation aims at assisting a naive user in modeling constraint problems. In this context, formal specification languages have been devised to express constraint problems in a manner similar to natural yet rigorous specifications that use a mixture of natural language and discrete mathematics. Yet, a gap remains between such languages and the natural language in which humans informally describe problems. This work aims to alleviate this issue by proposing a method for detecting constraints in natural language problem descriptions using a structured-output classifier. To evaluate the method, we develop an original annotated corpus which gathers 110 problem descriptions from several resources. Our results show significant accuracy with respect to metrics used in cognate tasks.
【Keywords】:
【Paper Link】 【Pages】:751-757
【Authors】: Jean-Marie Lagniez ; Emmanuel Lonca ; Pierre Marquis
【Abstract】: We present a new preprocessing technique for propositional model counting. This technique leverages definability, i.e., the ability to determine that some gates are implied by the input formula Σ. Such gates can be exploited to simplify Σ without modifying its number of models. Unlike previous techniques based on gate detection and replacement, gates do not need to be made explicit in our approach. Our preprocessing technique thus consists of two phases: computing a bipartition I, O of the variables of Σ where the variables from O are defined in Σ in terms of I, then eliminating some variables of O in Σ. Our experiments show the computational benefits which can be achieved by taking advantage of our preprocessing technique for model counting.
【Keywords】:
【Paper Link】 【Pages】:758-765
【Authors】: Jimmy H. M. Lee ; Zichen Zhu
【Abstract】: LexLeader, a state of the art static symmetry breaking method, adds a lex ordering constraint for each variable symmetry of the problem to select the lexicographically least solution. In practice, the same method can also be used for partial symmetry breaking by breaking only a given subset of symmetries. We propose a new total ordering, reflex, as basis of a new symmetry breaking constraint that collaborates well among themselves as well as with Precedence constraints, thereby breaking more composition symmetries in partial symmetry breaking. An efficient GAC filtering algorithm is presented for the reflex ordering constraint. We propose the ReflexLeader method, which is a variant of LexLeader using the reflex ordering instead, and give conditions when ReflexLeader is safe to combine with the Precedence and multiset ordering constraints. Extensive experimentations demonstrate empirically our claims and substantial advantages of the reflex ordering over the lex ordering in partial symmetry breaking.
【Keywords】:
【Paper Link】 【Pages】:766-772
【Authors】: Chu Min Li ; Felip Manyà ; Joan Ramon Soler
【Abstract】: We define a clause tableau calculus for MaxSAT, prove its soundness and completeness, and describe a tableau-based algorithm for MaxSAT. Given a multiset of clauses C, the algorithm computes both the minimum number of clauses that can be falsified in C, and an optimal assignment. We also describe how the algorithm can be extended to solve weighted MaxSAT and weighted partial MaxSAT.
【Keywords】:
【Paper Link】 【Pages】:773-779
【Authors】: Thierry Petit ; Lolita Petit
【Abstract】: In biology, the construction of plasmids is a routine technique, yet under-optimal, expensive and time-consuming. In this paper, we model the Plasmid Cloning Problem in constraint programing, in order to optimize the construction of plasmids. Our technique uses a new propagator for the AtMostNVector constraint. This constraint allows the design of strategies for constructing multiple plasmids at the same time. Our approach recommends the smallest number of different cloning steps, while selecting the most efficient steps. It provides optimal strategies for real instances in gene therapy for retinal blinding diseases.
【Keywords】:
【Paper Link】 【Pages】:780-786
【Authors】: Sam Snodgrass ; Santiago Ontañón
【Abstract】: Statistical models, such as Markov chains, have recently started to be studied for the purpose of Procedural Content Generation (PCG). A major problem with this approach is controlling the sampling process in order to obtain output satisfying some desired constraints. In this paper we present three approaches to constraining the content generated using multi-dimensional Markov chains: (1) a generate and test approach that simply resamples the content until the desired constraints are satisfied, (2) an approach that finds and resamples parts of the generated content that violate the constraints, and (3) an incremental method that checks for constraint violations during sampling. We test our approaches by generating maps for two classic video games, Super Mario Bros. and Kid Icarus.
【Keywords】:
【Paper Link】 【Pages】:787-795
【Authors】: Ruiwei Wang ; Wei Xia ; Roland H. C. Yap ; Zhanshan Li
【Abstract】: Maintaining Generalized Arc Consistency (GAC) during search is considered an efficient way to solve non-binary constraint satisfaction problems. Bit-based representations have been used effectively in Arc Consistency algorithms. We propose STRbit, a GAC algorithm, based on simple tabular reduction (STR) using an efficient bit vector support data structure. STRbit is extended to deal with compression of the underlying constraint with c-tuples. Experimental evaluation show our algorithms are faster than many algorithms (STR2, STR2-C, STR3, STR3-C and MDDc) across a variety of benchmarks except for problems with small tables where complex data structures do not payoff.
【Keywords】:
【Paper Link】 【Pages】:796-803
【Authors】: Ofra Amir ; Barbara J. Grosz ; Krzysztof Z. Gajos
【Abstract】: Complex collaborative activities such as treating patients, co-authoring documents and developing software are often characterized by teamwork that is loosely coupled and extends in time. To remain coordinated and avoid conflicts, team members need to identify dependencies between their activities — which though loosely coupled may interact — and share information appropriately. The loose-coupling of tasks increases the difficulty of identifying dependencies, with the result that team members often lack important information or are overwhelmed by irrelevant information. This paper formalizes a new multi-agent systems problem, Information Sharing in Loosely-Coupled Extended-Duration Teamwork (ISLET). It defines a new representation, Mutual Influence Potential Networks (MIP-Nets) and an algorithm, MIP-DOI, that uses this representation to determine the information that is most relevant to each team member. Importantly, because the extended duration of the teamwork precludes team-members developing complete plans in advance, the MIP-Nets approach, unlike prior work on information sharing, does not rely on a priori knowledge of a team's possible plans. Instead, it models collaboration patterns and dependencies among people and their activities based on team-members' interactions. Empirical evaluations show that this approach is able to learn collaboration patterns and identify relevant information to share with team members.
【Keywords】:
【Paper Link】 【Pages】:804-811
【Authors】: Ofra Amir ; Ece Kamar ; Andrey Kolobov ; Barbara J. Grosz
【Abstract】: Agents learning how to act in new environments can benefit from input from more experienced agents or humans. This paper studies interactive teaching strategies for identifying when a student can benefit from teacher-advice in a reinforcement learning framework. In student-teacher learning, a teacher agent can advise the student on which action to take. Prior work has considered heuristics for the teacher to choose advising opportunities. While these approaches effectively accelerate agent training, they assume that the teacher constantly monitors the student. This assumption may not be satisfied with human teachers, as people incur cognitive costs of monitoring and might not always pay attention. We propose strategies for a teacher and a student to jointly identify advising opportunities so that the teacher is not required to constantly monitor the student. Experimental results show that these approaches reduce the amount of attention required from the teacher compared to teacher-initiated strategies, while maintaining similar learning gains. The empirical evaluation also investigates the effect of the information communicated to the teacher and the quality of the student's initial policy on teaching outcomes.
【Keywords】:
【Paper Link】 【Pages】:812-818
【Authors】: Kai Chen ; Fangkai Yang ; Xiaoping Chen
【Abstract】: We propose a framework for a service robot to behave intelligently in domains that contain incomplete information, underspecified goals and dynamic change. Human robot interaction (HRI), sensing actions and physical actions are uniformly formalized in action language BC. An answer set solver is called to generate plans that guide the robot to acquire task-oriented knowledge and execute actions to achieve its goal, including interacting with human to gather information and sensing the environment to help motion planning. By continuously interpreting and grounding useful sensing information, robot is able to use contingent knowledge to adapt to unexpected changes and faults. We evaluate the approach on service robot KeJia that serves drink to guests, a testing benchmark for general-purpose service robot proposed by RoboCup@Home competition.
【Keywords】:
【Paper Link】 【Pages】:819-825
【Authors】: Shaofei Chen ; Tim Baarslag ; Dengji Zhao ; Jing Chen ; Lincheng Shen
【Abstract】: This paper studies a search problem involving a robot that is searching for a certain item in an uncertain environment (e.g., searching minerals on Moon) that allows only limited interaction with humans. The uncertainty of the environment comes from the rewards of undiscovered items and the availability of costly human help. The goal of the robot is to maximize the reward of the items found while minimizing the search costs. We show that this search problem is polynomially solvable with a novel integration of the human help, which has not been studied in the literature before. Furthermore, we empirically evaluate our solution with simulations and show that it significantly outperforms several benchmark approaches.
【Keywords】:
【Paper Link】 【Pages】:826-833
【Authors】: Matthew C. Gombolay ; Reed Jensen ; Jessica Stigile ; Sung-Hyun Son ; Julie A. Shah
【Abstract】: Coordinating agents to complete a set of tasks with intercoupled temporal and resource constraints is computationally challenging, yet human domain experts can solve these difficult scheduling problems using paradigms learned through years of apprenticeship. A process for manually codifying this domain knowledge within a computational framework is necessary to scale beyond the "single-expert, single-trainee apprenticeship model. However, human domain experts often have difficulty describing their decision-making processes, causing the codification of this knowledge to become laborious. We propose a new approach for capturing domain-expert heuristics through a pairwise ranking formulation. Our approach is model-free and does not require enumerating or iterating through a large state-space. We empirically demonstrate that this approach accurately learns multifaceted heuristics on both a synthetic data set incorporating job-shop scheduling and vehicle routing problems and a real-world data set consisting of demonstrations of experts solving a weapon-to-target assignment problem.
【Keywords】:
【Paper Link】 【Pages】:834-840
【Authors】: Eun-Sol Kim ; Kyoung-Woon On ; Byoung-Tak Zhang
【Abstract】: We explore the possibility of automatically constructing hierarchical schemas from low-level sensory data. Here we suggest a hierarchical event network to build the hierarchical schemas and describe a novel machine learning method to learn the network from the data. The traditional methods for describing schemas define the primitives and the relationships between them in advance. Therefore, it is difficult to adapt the constructed schemas in new situations. However, the proposed method constructs the schemas automatically from the data. Therefore, it has a novelty that the constructed schemas can be applied to new and unexpected situations flexibly. The key idea of constructing the hierarchical schema is selecting informative sensory data, integrating them sequentially and extracting high-level information. For the experiments, we collected sensory data using multiple wearable devices in restaurant situations. The experimental results demonstrate the real hierarchical schemas, which are probabilistic scripts and action primitives, constructed from the methods. Also, we show the constructed schemas can be used to predict the corresponding event to the low-level sensor data. Moreover, we show the prediction accuracy outperforms the conventional method significantly.
【Keywords】:
【Paper Link】 【Pages】:841-847
【Authors】: Ashish Kulkarni ; Pushpak Burange ; Ganesh Ramakrishnan
【Abstract】: We present an approach and a system that explores the application of interactive machine learning to a branching program-based boosting algorithm- Martingale Boosting. Typically, its performance is based on the ability of a learner to meet a fixed objective and does not account for preferences (e.g. low false positives) arising from an underlying classification problem. We use user preferences gathered on holdout data to guide the two-sided advantages of individual weak learners and tune them to meet these preferences. Extensive experiments show that while arbitrary preferences might be difficult to meet for a single classifier, a non-linear ensemble of classifiers as the one constructed by martingale boosting, performs better.
【Keywords】:
【Paper Link】 【Pages】:848-854
【Authors】: Thomas Moerland ; Joost Broekens ; Catholijn M. Jonker
【Abstract】: Social agents and robots will require both learning and emotional capabilities to successfully enter society. This paper connects both challenges, by studying models of emotion generation in sequential decision-making agents. Previous work in this field has focused on model-free reinforcement learning (RL). However, important emotions like hope and fear need anticipation, which requires a model and forward simulation. Taking inspiration from the psychological Belief-Desire Theory of Emotions (BDTE), our work specifies models of hope and fear based on best and worst forward traces. To efficiently estimate these traces, we integrate a well-known Monte Carlo Tree Search procedure (UCT) into a model based RL architecture. Test results in three known RL domains illustrate emotion dynamics, dependencies on policy and environmental stochasticity, and plausibility in individual Pacman game settings. Our models enable agents to naturally elicit hope and fear during learning, and moreover, explain what anticipated event caused this.
【Keywords】:
【Paper Link】 【Pages】:855-861
【Authors】: Zahra Nazari ; Jonathan Gratch
【Abstract】: Human and artificial negotiators must exchange information to find efficient negotiated agreements, but malicious actors could use deception to gain unfair advantage. The misrepresentation game is a game-theoretic formulation of how deceptive actors could gain disproportionate rewards while seeming honest and fair. Previous research proposed a solution to this game but this required restrictive assumptions that might render it inapplicable to real-world settings. Here we evaluate the formalism against a large corpus of human face-to-face negotiations. We confirm that the model captures how dishonest human negotiators win while seeming fair, even in unstructured negotiations. We also show that deceptive negotiators give-off signals of their malicious behavior, providing the opportunity for algorithms to detect and defeat this malicious tactic.
【Keywords】:
【Paper Link】 【Pages】:862-868
【Authors】: Stephanie Rosenthal ; Sai P. Selvaraj ; Manuela M. Veloso
【Abstract】: Autonomous mobile robots navigate in our spaces by planning and executing routes to destinations. When a mobile robot appears at a location, there is no clear way to understand what navigational path the robot planned and experienced just by looking at it. In this work, we address the generation of narrations of autonomous mobile robot navigation experiences. We contribute the concept of verbalization as a parallel to the well-studied concept of visualization. Through verbalizations, robots can describe through language what they experience, in particular in their paths. For every executed path, we consider many possible verbalizations that could be generated. We introduce the verbalization space that covers the variability of utterances that the robot may use to narrate its experience to different humans. We present an algorithm for segmenting a path and mapping each segment to an utterance, as a function of the desired point in the verbalization space, and demonstrate its application using our mobile service robot moving in our buildings. We believe our verbalization space and algorithm are applicable to different narrative aspects for many mobile robots, including autonomous cars.
【Keywords】:
【Paper Link】 【Pages】:869-877
【Authors】: Ngoc Cuong Truong ; Tim Baarslag ; Sarvapali D. Ramchurn ; Long Tran-Thanh
【Abstract】: We address the problem of recommending an appliance usage schedule to the homeowner which balances between maximizing total savings and maintaining sufficient user convenience. An important challenge within this problem is how to elicit the the user preferences with low intrusiveness, in order to identify new schedules with high cost savings, that still lies within the user's comfort zone. To tackle this problem we propose iDR, an interactive system for generating personalized appliance usage scheduling recommendations that maximize savings and convenience with minimal intrusiveness. In particular, our system learns when to stop interacting with the user during the preference elicitation process, in order to keep the bother cost (e.g., the amount of time the user spends, or the cognitive cost of interacting) minimal. We demonstrate through extensive empirical evaluation on real-world data that our approach improves savings by up to 35%, while maintaining a significantly lower bother cost, compared to state-of-the-art benchmarks.
【Keywords】:
【Paper Link】 【Pages】:878-885
【Authors】: Shqiponja Ahmetaj ; Magdalena Ortiz ; Mantas Simkus
【Abstract】: We consider instance queries mediated by an ontology expressed in the expressive DL ALCHIO with closed predicates. We observe that such queries are nonmonotonic and cannot be expressed in monotonic variants of Datalog, but a polynomial time translation into disjunctive Datalog extended with negation as failure is feasible. If no closed predicates are present — in the case of classical instance checking in ALCHIO — our translation yields a positive disjunctive Datalog program of polynomial size. To the best of our knowledge, this is the first polynomial time translation of an expressive (non-Horn) DL into disjunctive Datalog.
【Keywords】:
【Paper Link】 【Pages】:886-892
【Authors】: Mario Alviano ; Carmine Dodaro
【Abstract】: Clark's completion plays an important role in ASP computation:it discards unsupported models via unit resolution; hence, it improves the performance of ASP solvers, and at the same time it simplifies their implementation.In the disjunctive case, however, Clark's completion is usually preceded by another transformation known as shift, whose size is quadratic in general. A different approach is proposed in this paper: Clark's completion is extended to disjunctive programs without the need of intermediate program rewritings such as the shift.As in the non-disjunctive case, the new completion is linear in size, and discards unsupported models via unit resolution. Moreover, an ad-hoc propagator for supported model search is presented.
【Keywords】:
【Paper Link】 【Pages】:893-899
【Authors】: Antoine Amarilli ; Michael Benedikt ; Pierre Bourhis ; Michael Vanden Boom
【Abstract】: We consider entailment problems involving powerful constraint languages such as guarded existential rules, in which additional semantic restrictions are put on a set of distinguished relations. We consider restricting a relation to be transitive, restricting a relation to be the transitive closure of another relation, and restricting a relation to be a linear order. We give some natural generalizations of guardedness that allow inference to be decidable in each case, and isolate the complexity of the corresponding decision problems. Finally we show that slight changes in our conditions lead to undecidability.
【Keywords】:
【Paper Link】 【Pages】:900-906
【Authors】: Leila Amgoud ; Jonathan Ben-Naim
【Abstract】: This paper focuses on argumentation graphs whose nodes are arguments and edges represent supports, thus positive relations, between arguments. Furthermore, each argument has a weight reflecting its basic or intrinsic strength. For the sake of generality, the internal structure of arguments and the origin of arguments and their weights are unspecified. The paper tackles for the first time the question of evaluating the overall strengths of arguments in such graphs, thus of defining semantics for support graphs. It introduces a set of axioms that any semantics should satisfy. Then, it defines three semantics and evaluates them against the axioms.
【Keywords】:
【Paper Link】 【Pages】:907-914
【Authors】: Marcelo Arenas ; Jorge A. Baier ; Juan S. Navarro ; Sebastian Sardiña
【Abstract】: We propose a simple relaxation of Reiter's basic action theories, based on fluents without successor state axioms, that accommodates incompleteness beyond the initial database. We prove that fundamental results about basic action theories can be fully recovered and that the generalized framework allows for natural specifications of various forms of incomplete causal laws. We illustrate this by showing how the evolution of incomplete databases, guarded action theories, and non-deterministic actions can be conveniently specified.
【Keywords】:
【Paper Link】 【Pages】:915-921
【Authors】: Evgenii Balai ; Michael Gelfond
【Abstract】: The paper investigates the relationship between knowledge representation languages P-log and LPMLN designed for representing and reasoning with logic and probability. We give a translation from an important subset of LPMLN to P-log which preserves probabilistic functions defined by LPMLN programs and complements recent research by the authors of LPMLN where they give a similar translation from a subset of P-log to their language. This work sheds light on the different ways to treat inconsistency in both languages.
【Keywords】:
【Paper Link】 【Pages】:922-928
【Authors】: Bita Banihashemi ; Giuseppe De Giacomo ; Yves Lespérance
【Abstract】: Agent supervision is a form of control / customization where a supervisor restricts the behavior of an agent to enforce certain requirements, while leaving the agent as much autonomy as possible. In this work, we investigate supervision of an agent that may acquire new knowledge about her environment during execution, for example, by sensing. Thus we consider an agent's online executions, where, as she executes the program, at each time point she must make decisions on what to do next based on what her current knowledge is. This is done in a setting based on the situation calculus and a variant of the ConGolog programming language. The main results of this paper are (i) a formalization of the online maximally permissive supervisor, (ii) a sound and complete technique for execution of the agent as constrained by such a supervisor, and (iii)a new type of lookahead search construct that ensures nonblockingness over such online executions.
【Keywords】:
【Paper Link】 【Pages】:929-935
【Authors】: Harald Beck ; Minh Dao-Tran ; Thomas Eiter
【Abstract】: The emerging research field of stream reasoning faces the challenging trade-off between expressiveness of query programs and data throughput. For optimizing programs methods are needed to tell whether two programs are equivalent. Towards providing practical reasoning techniques on streams, we consider LARS programs, which is a powerful extension of Answer Set Programming (ASP) for stream reasoning that supports windows on streams for discarding information.We define different notions of equivalence between such programs and give semantic characterizations in terms of models. We show how a practically relevant fragment can be alternatively captured usingHere-and-There models, yielding an extension of equilibrium semantics of ASP to this class of programs. Finally, we characterize the computational complexity of deciding the considered equival encerelations.
【Keywords】:
【Paper Link】 【Pages】:936-942
【Authors】: Christoph Benzmüller ; Bruno Woltzenlogel Paleo
【Abstract】: This paper discusses the discovery of the inconsistency in Gödel's ontological argument as a success story for artificial intelligence. Despite the popularity of the argument since the appearance of Gödel's manuscript in the early 1970s, the inconsistency of the axioms used in the argument remained unnoticed until 2013, when it was detected automatically by the higher-order theorem prover Leo-II. Understanding and verifying the refutation generated by the prover turned out to be a time-consuming task. Its completion, as reported here, required the reconstruction of the refutation in the Isabelle proof assistant, and it also led to a novel and more efficient way of automating higher-order modal logic S5 with a universal accessibility relation. Furthermore, the development of an improved syntactical hiding for the utilized logic embedding technique allows the refutation to be presented in a human-friendly way, suitable for non experts in the technicalities of higher-order theorem proving. This brings us a step closer to wider adoption of logic-based artificial intelligence tools by philosophers.
【Keywords】:
【Paper Link】 【Pages】:943-949
【Authors】: Gerald Berger ; Andreas Pieris
【Abstract】: Ontology-based data access is concerned with the problem of querying incomplete data sources in the presence of an ontology. A key notion in this setting is that of ontology-mediated query, which is a database query coupled with an ontology. An interesting issue is whether the answer to an ontology-mediated query can be computed by parallelizing it over the connected components of the database, i.e., whether the query distributes over components. This allows us to evaluate the query in a distributed and coordination-free manner. We investigate distribution over components for classes of ontology-mediated queries where the database query is a conjunctive query and the ontology is formulated using existential rules. For each such class, we syntactically characterize its fragment that distributes over components, and we study the problem of deciding whether a query distributes over components.
【Keywords】:
【Paper Link】 【Pages】:950-956
【Authors】: Matteo Bertello ; Nicola Gigante ; Angelo Montanari ; Mark Reynolds
【Abstract】: The paper presents Leviathan, an LTL satisfiability checking tool based on a novel one-pass, tree-like tableau system, which is way simpler than existing solutions. Despite the simplicity of the algorithm, the tool has performance comparable in speed and memory consumption with other tools on a number of standard benchmark sets, and, in various cases, it outperforms the other tableau-based tools.
【Keywords】:
【Paper Link】 【Pages】:957-964
【Authors】: Meghyn Bienvenu ; Camille Bourgaux ; François Goasdoué
【Abstract】: We consider the problem of query-driven repairing of inconsistent DL-Lite knowledge bases: query answers are computed under inconsistency-tolerant semantics, and the user provides feedback about which answers are erroneous or missing. The aim is to find a set of ABox modifications (deletions and additions), called a repair plan, that addresses as many of the defects as possible. After formalizing this problem and introducing different notions of optimality, we investigate the computational complexity of reasoning about optimal repair plans and propose interactive algorithms for computing such plans. For deletion-only repair plans, we also present a prototype implementation of the core components of the algorithm.
【Keywords】:
【Paper Link】 【Pages】:965-971
【Authors】: Meghyn Bienvenu ; Peter Hansen ; Carsten Lutz ; Frank Wolter
【Abstract】: We study FO-rewritability of conjunctive queries in the presence of ontologies formulated in a description logic between EL and Horn-SHIF, along with related query containment problems. Apart from providing characterizations, we establish complexity results ranging from ExpTime via NExpTime to 2ExpTime, pointing out several interesting effects. In particular, FO-rewriting is more complex for conjunctive queries than for atomic queries when inverse roles are present, but not otherwise.
【Keywords】:
【Paper Link】 【Pages】:972-978
【Authors】: Benjamin Bittner ; Marco Bozzano ; Alessandro Cimatti
【Abstract】: Timed Failure Propagation Graphs (TFPGs) are used in the design of safety-critical systems as a way of modeling failure propagation, and to evaluate and implement diagnostic systems. TFPGs are mostly produced manually, from a given dynamic system of greater complexity. In this paper we present a technique to automate the construction of TFPGs. It takes as input a set of failure mode and discrepancy nodes and builds the graph on top of them, based on an exhaustive analysis of all system behaviors. The result is a TFPG that accurately represents the sequences of failures and their effects as they appear in the system model. The proposed approach has been implemented on top of state-of-the-art symbolic model-checking techniques, and thoroughly evaluated on a number of synthetic and industrial benchmarks.
【Keywords】:
【Paper Link】 【Pages】:979-986
【Authors】: Bernhard Bliem ; Benjamin Kaufmann ; Torsten Schaub ; Stefan Woltran
【Abstract】: Answer Set Programming (ASP) has recently been employed to specify and run dynamic programming (DP) algorithms on tree decompositions, a central approach in the field of parameterized complexity, which aims at solving hard problems efficiently for instances of certain structure. This ASP-based method followed the standard DP approach where tables are computed in a bottom-up fashion, yielding good results for several counting or enumeration problems. However, for optimization problems this approach lacks the possibility to report solutions before the optimum is found, and for search problems it often computes a lot of unnecessary rows. In this paper, we present a novel ASP-based system allowing for "lazy" DP, which utilizes recent multi-shot ASP technology. Preliminary experimental results show that this approach not only yields better performance for search problems, but also outperforms some state-of-the-art ASP encodings for optimization problems in terms of anytime computation, i.e., measuring the quality of the best solution after a certain timeout.
【Keywords】:
【Paper Link】 【Pages】:987-993
【Authors】: Richard Booth ; Jake Chandler
【Abstract】: The field of iterated belief change has focused mainly on revision, with the other main operator of AGM belief change theory, i.e., contraction receiving relatively little attention. In this paper we extend the Harper Identity from single-step change to define iterated contraction in terms of iterated revision. Specifically, just as the Harper Identity provides a recipe for defining the belief set resulting from contracting A in terms of (i) the initial belief set and (ii) the belief set resulting from revision by not-A, we look at ways to define the plausibility ordering over worlds resulting from contracting A in terms of (iii) the initial plausibility ordering, and (iv) the plausibility ordering resulting from revision by not-A. After noting that the most straightforward such extension leads to a trivialisation of the space of permissible orderings, we provide a family of operators for combining plausibility orderings that avoid such a result. These operators are characterised in our domain of interest by a pair of intuitively compelling properties, which turn out to enable the derivation of a number of iterated contraction postulates from postulates for iterated revision. We finish by observing that a salient member of this family allows for the derivation of counterparts for contraction of some well known iterated revision operators, as well as for defining new iterated contraction operators.
【Keywords】:
【Paper Link】 【Pages】:994-1000
【Authors】: Stefan Borgwardt ; Bettina Fazzinga ; Thomas Lukasiewicz ; Akanksha Shrivastava ; Oana Tifrea-Marciuska
【Abstract】: In this paper, we explore how ontological knowledge expressed via existential rules can be combined with possibilistic networks (i) to represent qualitative preferences along with domain knowledge, and (ii) to realize preference-based answering of conjunctive queries (CQs). We call these combinations ontological possibilistic networks (OP-nets). We define skyline and k-rank answers to CQs under preferences and provide complexity (including data tractability) results for deciding consistency and CQ skyline membership for OP-nets. We show that our formalism has a lower complexity than a similar existing formalism.
【Keywords】:
【Paper Link】 【Pages】:1001-1007
【Authors】: Elena Botoeva ; Carsten Lutz ; Vladislav Ryzhikov ; Frank Wolter ; Michael Zakharyaschev
【Abstract】: We investigate the problem whether two ALC knowledge bases are indistinguishable by queries over a given vocabulary. We give model-theoretic criteria and prove that this problem is undecidable for conjunctive queries (CQs) but decidable in 2EXPTIME for unions of rooted CQs. We also consider the problem whether two ALC TBoxes give the same answers to any query in a given vocabulary over all ABoxes, and show that for CQs this problem is undecidable, too, but becomes decidable and 2EXPTIME-complete in Horn-ALC, and even EXPTIME-complete in Horn-ALC when restricted to (unions of) rooted CQs.
【Keywords】:
【Paper Link】 【Pages】:1008-1014
【Authors】: Simone Bova ; Florent Capelli ; Stefan Mengel ; Friedrich Slivovsky
【Abstract】: Choosing a language for knowledge representation and reasoning involves a trade-off between two competing desiderata: succinctness (the encoding should be small) and tractability (the language should support efficient reasoning algorithms). The area of knowledge compilation is devoted to the systematic study of representation languages along these two dimensions — in particular, it aims to determine the relative succinctness of languages. Showing that one language is more succinct than another typically involves proving a nontrivial lower bound on the encoding size of a carefully chosen function, and the corresponding arguments increase in difficulty with the succinctness of the target language. In this paper, we introduce a general technique for obtaining lower bounds on Decomposable Negation Normal Form (DNNFs), one of the most widely studied and succinct representation languages, by relating the size of DNNFs to multi-partition communication complexity. This allows us to directly translate lower bounds from the communication complexity literature into lower bounds on the size of DNNF representations. We use this approach to prove exponential separations of DNNFs from deterministic DNNFs and of CNF formulas from DNNFs.
【Keywords】:
【Paper Link】 【Pages】:1015-1021
【Authors】: Pedro Cabalar ; Roland Kaminski ; Max Ostrowski ; Torsten Schaub
【Abstract】: We introduce the logic of Here-and-There with Constraints in order to capture constraint theories in the non-monotonic setting known from Answer Set Programming (ASP). This allows for assigning default values to constraint variables or to leave them undefined. Also, it provides us with a semantic framework integrating ASP and Constraint Processing in a uniform way. We put some emphasis on logic programs dealing with linear constraints on integer variables, where we further introduce a directional assignment operator. We elaborate upon the formal relation and implementation of these programs in terms of Constraint ASP, sketching an existing system.
【Keywords】:
【Paper Link】 【Pages】:1022-1029
【Authors】: Diego Calvanese ; Marco Montali ; Fabio Patrizi ; Michele Stawowy
【Abstract】: We study plan synthesis for a variant of Knowledge and Action Bases (KABs), a rich, dynamic framework, where states are description logic (DL) knowledge bases (KBs) whose extensional part is manipulated by actions that possibly introduce new objects from an infinite domain. We show that plan existence over KABs is undecidable even under severe restrictions. We then focus on state-bounded KABs, a class for which plan existence is decidable, and provide sound and complete plan synthesis algorithms, which combine techniques based on standard planning, DL query answering, and finite-state abstraction. All results hold for any DL with decidable query answering. We finally show that for lightweight DLs, plan synthesis can be compiled into standard ADL planning.
【Keywords】:
【Paper Link】 【Pages】:1030-1036
【Authors】: Tristan Charrier ; Bastien Maubert ; François Schwarzentruber
【Abstract】: Epistemic planning is a variant of automated planning in the framework of dynamic epistemic logic. In recent works, the epistemic planning problem has been proved to be undecidable when preconditions of events can be epistemic formulas of arbitrary complexity, and in particular arbitrary modal depth. It is known however that when preconditions are propositional (and there are no postconditions), the problem is between PSPACE and EXPSPACE. In this work we bring two new pieces to the picture. First, we prove that the epistemic planning problem with propositional preconditions and without postconditions is in PSPACE, and is thus PSPACE-complete. Second, we prove that very simple epistemic preconditions are enough to make the epistemic planning problem undecidable: preconditions of modal depth at most two suffice.
【Keywords】:
【Paper Link】 【Pages】:1037-1043
【Authors】: Giuseppe De Giacomo ; Aniello Murano ; Sasha Rubin ; Antonio Di Stasio
【Abstract】: We study a generalized form of planning under partial observability, in which we have multiple, possibly infinitely many, planning domains with the same actions and observations, and goals expressed over observations, which are possibly temporally extended. By building on work on two-player (non-probabilistic) games with imperfect information in the Formal Methods literature, we devise a general technique, generalizing the belief-state construction, to remove partial observability. This reduces the planning problem to a game of perfect information with a tight correspondence between plans and strategies. Then we instantiate the technique and solve some generalized-planning problems.
【Keywords】:
【Paper Link】 【Pages】:1044-1050
【Authors】: Giuseppe De Giacomo ; Moshe Y. Vardi
【Abstract】: In this paper, we study synthesis under partial observability for logical specifications over finite traces expressed in LTLf/LDLf. This form of synthesis can be seen as a generalization of planning under partial observability in nondeterministic domains, which is known to be 2EXPTIME-complete. We start by showing that the usual "belief-state construction" used in planning under partial observability works also for general LTLf/LDLf synthesis, though with a jump in computational complexity from 2EXPTIME to 3EXPTIME. Then we show that the belief-state construction can be avoided in favor of a direct automata construction which exploits projection to hide unobservable propositions. This allow us to prove that the problem remains 2EXPTIME-complete. The new synthesis technique proposed is effective and readily implementable.
【Keywords】:
【Paper Link】 【Pages】:1051-1057
【Authors】: Paul E. Dunne ; Christof Spanring ; Thomas Linsbichler ; Stefan Woltran
【Abstract】: Understanding the relation between different semantics in abstract argumentation is an important issue, not least since such semantics capture the basic ingredients of different approaches to nonmonotonic reasoning. The question we are interested in relates two semantics as follows: What are the necessary and sufficient conditions, such that we can decide, for any two sets of extensions, whether there exists an argumentation framework which has exactly the first extension set under one semantics, and the second extension set under the other semantics. We investigate in total nine argumentation semantics and give a nearly complete landscape of exact characterizations. As we shall argue, such results not only give an account on the independency between semantics, but might also prove useful in argumentation systems by providing guidelines for how to prune the search space.
【Keywords】:
【Paper Link】 【Pages】:1058-1065
【Authors】: Thomas Eiter ; Tobias Kaminski ; Christoph Redl ; Antonius Weinzierl
【Abstract】: Answer Set Programming (ASP) is a well-known problem solving approach based on nonmonotonic logic programs and efficient solvers. HEX-programs extend ASP with external atoms for access to arbitrary external information. In this work, we extend the evaluation principles of external atoms to partial assignments, lift nogood learning to this setting, and introduce a variant of nogood minimization. This enables external sources to guide the search for answer sets akin to theory propagation. Our benchmark experiments demonstrate a clear improvement in efficiency over the state-of-the-art HEX-solver.
【Keywords】:
【Paper Link】 【Pages】:1066-1073
【Authors】: Liangda Fang ; Yongmei Liu ; Hans van Ditmarsch
【Abstract】: In the past decades, forgetting has been investigated for many logics and has found many applications in knowledge representation and reasoning. However, forgetting in multi-agent modal logics has largely been unexplored. In this paper, we study forgetting in multi-agent modal logics. We adopt the semantic definition of existential bisimulation quantifiers as that of forgetting. We propose a syntactical way of performing forgetting based on the canonical formulas of modal logics introduced by Moss. We show that the result of forgetting a propositional atom from a satisfiable canonical formula can be computed by simply substituting the literals of the atom with T. Thus we show that Kn, Dn, Tn, K45n, KD45n and S5n are closed under forgetting, and hence have uniform interpolation.
【Keywords】:
【Paper Link】 【Pages】:1074-1080
【Authors】: Xiaoyu Ge ; Jae Hee Lee ; Jochen Renz ; Peng Zhang
【Abstract】: The capability to predict changes of spatial regions is important for an intelligent system that interacts with the physical world. For example, in a disaster management scenario, predicting potentially endangered areas and inferring safe zones is essential for planning evacuations and countermeasures. Existing approaches usually predict such spatial changes by simulating the physical world based on specific models. Thus, these simulation-based methods will not be able to provide reliable predictions when the scenario is not similar to any of the models in use or when the input parameters are incomplete. In this paper, we present a prediction approach that overcomes the aforementioned problem by using a more general model and by analysing the trend of the spatial changes. The method is also flexible to adopt to new observations and to adapt its prediction to new situations.
【Keywords】:
【Paper Link】 【Pages】:1081-1087
【Authors】: Anne-Marie George ; Nic Wilson ; Barry O'Sullivan
【Abstract】: In this paper, we construct and compare algorithmic approaches to solve the Preference Consistency Problem for preference statements based on hierarchical models. Instances of this problem contain a set of preference statements that are direct comparisons (strict and non-strict) between some alternatives, and a set of evaluation functions by which all alternatives can be rated. An instance is consistent based on hierarchical preference models, if there exists an hierarchical model on the evaluation functions that induces an order relation on the alternatives by which all relations given by the preference statements are satisfied. Deciding if an instance is consistent is known to be NP-complete for hierarchical models. We develop three approaches to solve this decision problem. The first involves a Mixed Integer Linear Programming (MILP) formulation, the other two are recursive algorithms that are based on properties of the problem by which the search space can be pruned. Our experiments on synthetic data show that the recursive algorithms are faster than solving the MILP formulation and that the ratio between the running times increases extremely quickly.
【Keywords】:
【Paper Link】 【Pages】:1088-1094
【Authors】: Maciej Grabon ; Jakub Michaliszyn ; Jan Otop ; Piotr Wieczorek
【Abstract】: We propose a query language LARE for graphs whose edges are labelled by a finite alphabet and nodes store unbounded data values. LARE is based on a variant of regular expressions with memory. Queries of LARE can compare quantities of memorised graph nodes and their neighbourhoods. These features allow us to express a number of natural properties, while keeping the data complexity (with a query fixed) in non-deterministic logarithmic space. We establish an algorithm that works efficiently not only with LARE, but also with a wider language defined using effective relational conditions, another formalism we propose.
【Keywords】:
【Paper Link】 【Pages】:1095-1101
【Authors】: Éric Grégoire ; Sébastien Konieczny ; Jean-Marie Lagniez
【Abstract】: Computing a consensus is a key task in various AI areas, ranging from belief fusion, social choice,negotiation, etc. In this work, we define consensus operators as functions that deliver parts of the set-theoretical union of the information sources (inpropositional logic) to be reconciled, such that no source is logically contradicted. We also investigate different notions of maximality related to these consensuses. From a computational point of view, we propose a generic problem transformation that leads to a method that proves experimentally efficient very often, even for large conflicting sources to be reconciled.
【Keywords】:
【Paper Link】 【Pages】:1102-1108
【Authors】: Víctor Gutiérrez-Basulto ; Jean Christoph Jung ; Roman Kontchakov
【Abstract】: We study access to temporal data with TEL, a temporal extension of the tractable description logic EL. Our aim is to establish a clear computational complexity landscape for the atomic query answering problem, in terms of both data and combined complexity. Atomic queries in full TEL turn out to be undecidable even in data complexity. Motivated by the negative result, we identify well-behaved yet expressive fragments of TEL. Our main contributions are a semantic and sufficient syntactic conditions for decidability and three orthogonal tractable fragments, which are based on restricted use of rigid roles, temporal operators, and novel acyclicity conditions on the ontologies.
【Keywords】:
【Paper Link】 【Pages】:1109-1115
【Authors】: Adrian Haret ; Jean-Guy Mailly ; Stefan Woltran
【Abstract】: Understanding the behavior of belief change operators for fragments of classical logic has received increasing interest over the last years. Results in this direction are mainly concerned with adapting representation theorems. However, fragment-driven belief change also leads to novel research questions. In this paper we propose the concept of belief distribution, which can be understood as the reverse task of merging. More specifically, we are interested in the following question: given an arbitrary knowledge base K and some merging operator Δ, can we find a profile E and a constraint μ, both from a given fragment of classical logic, such that Δμ(E) yields a result equivalent to K? In other words, we are interested in seeing if K can be distributed into knowledge bases of simpler structure, such that the task of merging allows for a reconstruction of the original knowledge. Our initial results show that merging based on drastic distance allows for an easy distribution of knowledge, while the power of distribution for operators based on Hamming distance relies heavily on the fragment of choice.
【Keywords】:
【Paper Link】 【Pages】:1116-1122
【Authors】: Andreas Herzig ; Emiliano Lorini ; Faustine Maffre ; François Schwarzentruber
【Abstract】: We analyse epistemic boolean games ina computationally grounded dynamic epistemic logic. The agents' knowledge is determined by what they see, including higher-order visibility: agents may observe whether another agent observes an atom or not. The agents' actions consist in modifying the truth values of atoms. We provide an axiomatisation of the logic, establish that the model checking problem is in PSPACE, and show how one can reason about equilibria in epistemic boolean games.
【Keywords】:
【Paper Link】 【Pages】:1123-1129
【Authors】: Xiaowei Huang ; Ji Ruan ; Qingliang Chen ; Kaile Su
【Abstract】: Social norms are powerful formalism in coordinating autonomous agents' behaviour to achieve certain objectives. In this paper, we propose a dynamic normative system to enable the reasoning of the changes of norms under different circumstances, which cannot be done in the existing static normative systems. We study two important problems (norm synthesis and norm recognition) related to the autonomy of the entire system and the agents, and characterise the computational complexities of solving these problems.
【Keywords】:
【Paper Link】 【Pages】:1130-1137
【Authors】: Jianmin Ji ; Hai Wan ; Kewen Wang ; Zhe Wang ; Chuhan Zhang ; Jiangtao Xu
【Abstract】: A disjunctive logic program under the answer set semantics can be equivalently translated to a normal logic program by the shifting transformation, if the program is head-cycle-free. In this paper, we provide an answer-set-preserving rewriting of a general disjunctive program to a normal program by first applying the unfolding transformation on atoms that prevent the program from being head-cycle-free, then shifting the resulting program. Different from other transformations that eliminate disjunctions in answer set programming, the new rewriting is efficient for "almost" head-cycle-free programs, i.e., programs that have only a few atoms that prevent them to be head-cycle-free. Based on the new rewriting, we provide an anytime algorithm to compute answer sets of a disjunctive program by calling solvers for normal logic programs. The experiment shows that the algorithm is efficient for some disjunctive programs. We also extend the rewriting to non-ground answer set programs on finite structures.
【Keywords】:
【Paper Link】 【Pages】:1138-1144
【Authors】: Guifei Jiang ; Dongmo Zhang ; Laurent Perrussel ; Heng Zhang
【Abstract】: This paper proposes a logical framework for representing and reasoning about imperfect information games. We first extend the game description language (GDL) with the standard epistemic operators and provide it with a semantics based on the epistemic state transition model. We then demonstrate how to use the language to represent the rules of an imperfect information game and formalize its epistemic properties. We also show how to use the framework to reason about player's own as well as other players' knowledge during game playing. Finally we prove that the model-checking problem of the framework is in Δ2p, which is the lowest among the existing similar frameworks, even though its lower bound is Θ2p. These results indicate that the framework makes a good balance between expressive power and computational efficiency.
【Keywords】:
【Paper Link】 【Pages】:1145-1152
【Authors】: Daniel Khashabi ; Tushar Khot ; Ashish Sabharwal ; Peter Clark ; Oren Etzioni ; Dan Roth
【Abstract】: Answering science questions posed in natural language is an important AI challenge. Answering such questions often requires non-trivial inference and knowledge that goes beyond factoid retrieval. Yet, most systems for this task are based on relatively shallow Information Retrieval (IR) and statistical correlation techniques operating on large unstructured corpora. We propose a structured inference system for this task, formulated as an Integer Linear Program (ILP), that answers natural language questions using a semi-structured knowledge base derived from text, including questions requiring multi-step inference and a combination of multiple facts. On a dataset of real, unseen science questions, our system significantly outperforms (+14%) the best previous attempt at structured reasoning for this task, which used Markov Logic Networks (MLNs). It also improves upon a previous ILP formulation by 17.7%. When combined with unstructured inference methods, the ILP system significantly boosts overall performance (+10%). Finally, we show our approach is substantially more robust to a simple answer perturbation compared to statistical correlation methods.
【Keywords】:
【Paper Link】 【Pages】:1153-1159
【Authors】: Boris Konev ; Carsten Lutz ; Frank Wolter ; Michael Zakharyaschev
【Abstract】: We investigate the problem of conservative rewritability of a TBox T in a description logic L into a TBox T' in a weaker description logic L'. We focus on model-conservative rewritability (T' entails T and all models of T are expandable to models of T'), subsumption-conservative rewritability (T' entails T and all subsumptions in the signature of T entailed by T' are entailed by T), and standard description logics between ALC and ALCQI. We give model-theoretic characterizations of conservative rewritability via bisimulations, inverse p-morphisms and generated subinterpretations, and use them to obtain a few rewriting algorithms and complexity results for deciding rewritability.
【Keywords】:
【Paper Link】 【Pages】:1160-1166
【Authors】: Roman Kontchakov ; Laura Pandolfo ; Luca Pulina ; Vladislav Ryzhikov ; Michael Zakharyaschev
【Abstract】: We design an extension datalogHS of datalog with hyperrectangle generalisations of Halpern-Shoham's modal operators on intervals and a corresponding query language. We prove that, over n-dimensional spaces comprised of Z and R, finding certain answers to datalogHS queries can be reduced to standard datalog query answering. We present experimental results showing the expressivity and efficiency of datalogHS on historical data.
【Keywords】:
【Paper Link】 【Pages】:1167-1173
【Authors】: Ondrej Kuzelka ; Jesse Davis ; Steven Schockaert
【Abstract】: We introduce a setting for learning possibilistic logic theories from defaults of the form "if alpha then typically beta". We first analyse this problem from the point of view of machine learning theory, determining the VC dimension of possibilistic stratifications as well as the complexity of the associated learning problems, after which we present a heuristic learning algorithm that can easily scale to thousands of defaults. An important property of our approach is that it is inherently able to handle noisy and conflicting sets of defaults. Among others, this allows us to learn possibilistic logic theories from crowdsourced data and to approximate propositional Markov logic networks using heuristic MAP solvers. We present experimental results that demonstrate the effectiveness of this approach.
【Keywords】:
【Paper Link】 【Pages】:1174-1180
【Authors】: Maurizio Lenzerini ; Lorenzo Lepore ; Antonella Poggi
【Abstract】: Hi(OWL 2 QL) is a new ontology language with the OWL2QL syntax and a specific semantics designed to support metamodeling and meta-querying. In this paper we investigate the problem of answering metaqueries in Hi(OWL 2 QL), which are unions of conjunctive queries with both ABox and TBox atoms. We first focus on a specific class of ontologies, called TBox-complete, where there is no uncertainty about TBox axioms, and show that query answering in this case has the same complexity (both data and combined) as in OWL2QL. We then move to general ontologies and show that answering metaqueries is coNP-complete with respect to ontology complexity, Π2p-complete with respect to combined complexity, and remains AC0 with respect to ABox complexity. Finally, we present an optimized query answering algorithm that can be used for TBox-complete ontologies.
【Keywords】:
【Paper Link】 【Pages】:1181-1187
【Authors】: Yuliya Lierler ; Benjamin Susman
【Abstract】: Constraint answer set programming is a promising research direction that integrates answer set programming with constraint processing. It is often informally related to the field of Satisfiability Modulo Theories. Yet, the exact formal link is obscured as the terminology and concepts used in these two research areas differ. In this paper, we make the link between these two areas precise.
【Keywords】:
【Paper Link】 【Pages】:1188-1194
【Authors】: Xin Liu ; Yong Liu ; Xiaoli Li
【Abstract】: Conventional location recommendation models rely on users' visit history, geographical influence, temporal influence, etc., to infer users' preferences for locations. However, systematically modeling a location's context (i.e., the set of locations visited before or after this location) is relatively unexplored. In this paper, by leveraging the Skip-gram model, we learn the latent representation for a location to capture the influence of its context. A pair-wise ranking loss that considers the confidences of observed user preferences for locations is then proposed to learn users' latent representations for personalized top-N location recommendations. Moreover, we also extend our model by taking into account temporal influence. Stochastic gradient descent based optimization algorithms are developed to fit the models. We conduct comprehensive experiments over four real datasets. Experimental results demonstrate that our approach significantly outperforms the state-of-the-art location recommendation methods.
【Keywords】:
【Paper Link】 【Pages】:1195-1201
【Authors】: Yao Liu ; Zhenhua Duan ; Cong Tian
【Abstract】: In this paper, we study an expressive fragment, namely Gmu, of linear time mu-calculus as a high-level goal specification language. We define Goal Progression Form (GPF) for Gmu formulas and show that every closed formula can be transformed into this form. Based on GPF, we present the notion of Goal Progression Form Graph (GPG) which can be used to describe models of a formula. Further, we propose a simple and intuitive GPG-based decision procedure for checking satisfiability of Gmu formulas which has the same time complexity as the decision problem of Linear Temporal Logic (LTL). However, Gmu is able to express a wider variety of temporal goals compared with LTL.
【Keywords】:
【Paper Link】 【Pages】:1202-1208
【Authors】: Zhiguo Long ; Michael Sioutis ; Sanjiang Li
【Abstract】: We propose a new algorithm called DPC+ to enforce partial path consistency (PPC) on qualitative constraint networks. PPC restricts path consistency (PC) to a triangulation of the underlying constraint graph of a network. As PPC retains the sparseness of a constraint graph, it can make reasoning tasks such as consistency checking and minimal labelling of large qualitative constraint networks much easier to tackle than PC. For qualitative constraint networks defined over any distributive subalgebra of well-known spatio-temporal calculi, such as the Region Connection Calculus and the Interval Algebra, we show that DPC+ can achieve PPC very fast. Indeed, the algorithm enforces PPC on a qualitative constraint network by processing each triangle in a triangulation of its underlying constraint graph at most three times. Our experiments demonstrate significant improvements of DPC+ over the state-of-the-art PPC enforcing algorithm.
【Keywords】:
【Paper Link】 【Pages】:1209-1215
【Authors】: Tim Miller ; Christian J. Muise
【Abstract】: Reasoning about the nested beliefs of others is important in many multi-agent scenarios. While epistemic and doxastic logics lay a solid groundwork to approach such reasoning, the computational complexity of these logics is often too high for many tasks. Proper Epistemic Knowledge Bases (PEKBs) enforce two syntactic restrictions on formulae to obtain efficient querying: both disjunction and infinitely long nestings of modal operators are not permitted. PEKBs can be compiled, in exponential time, to a prime implicate formula that can be queried in polynomial time, while more recently, it was shown that consistent PEKBs had certain logical properties that meant this compilation was unnecessary, while still retaining polynomial-time querying. In this paper, we present a belief update mechanism for PEKBs that ensures the knowledge base remains consistent when new beliefs are added. This is achieved by first erasing any formulae that contradict these new beliefs. We show that this update mechanism can be computed in polynomial time, and we assess it against the well-known KM postulates for belief update.
【Keywords】:
【Paper Link】 【Pages】:1216-1222
【Authors】: Andreas Niskanen ; Johannes Peter Wallner ; Matti Järvisalo
【Abstract】: We present complexity results and algorithms for optimal status enforcement in abstract argumentation. Status enforcement is the task of adjusting a given argumentation framework (AF) to support given positive and negative argument statuses, i.e., to accept and reject specific arguments. We study optimal status enforcement as the problem of finding a structurally closest AF supporting given argument statuses. We establish complexity results for optimal status enforcement under several central AF semantics, develop constraint-based algorithms for NP and second-level complete variants of the problem, and empirically evaluate the procedures.
【Keywords】:
【Paper Link】 【Pages】:1223-1229
【Authors】: Alexandre Niveau ; Bruno Zanuttini
【Abstract】: We investigate efficient representations of subjective formulas in the modal logic of knowledge, S5, and more generally of sets of sets of propositional assignments. One motivation for this study is contingent planning, for which many approaches use operations on such formulas, and can clearly take advantage of efficient representations. We study the language S5-DNF introduced by Bienvenu et al., and a natural variant of it that uses Binary Decision Diagrams at the propositional level. We also introduce an alternative language, called Epistemic Splitting Diagrams, which provides more compact representations. We compare all three languages from the complexity-theoretic viewpoint of knowledge compilation and also through experiments. Our work sheds light on the pros and cons of each representation in both theory and practice.
【Keywords】:
【Paper Link】 【Pages】:1230-1236
【Authors】: Sebastian Rudolph ; Michaël Thomazo
【Abstract】: Computational and model-theoretic properties of logical languages constitute a central field of research in logic-based knowledge representation. Datalog is a very popular formalism, a de-facto standard for expressing and querying knowledge. Diverse results exist regarding the expressivity of Datalog and its extension by input negation (semipositive Datalog) and/or a linear order (order-invariant Datalog). When classifying the expressivity of logical formalisms by their model-theoretic properties, a very natural and prominent such property is preservation under homomorphisms. This paper solves the remaining open questions needed to arrive at a complete picture regarding the interrelationships between the class of homomorphism-closed queries and the query classes related to the four versions of Datalog. Most notably, we exhibit a query that is both homomorphism-closed and computable in polynomial time but cannot be expressed in order-invariant Datalog.
【Keywords】:
【Paper Link】 【Pages】:1237-1243
【Authors】: Nicolas Schwind ; Katsumi Inoue ; Gauvain Bourgne ; Sébastien Konieczny ; Pierre Marquis
【Abstract】: We consider the problem of belief propagation in a network of communicating agents, modeled in the recently introduced Belief Revision Game (BRG) framework. In this setting, each agent expresses her belief through a propositional formula and revises her own belief at each step by considering the beliefs of her acquaintances, using belief change tools. In this paper, we investigate the extent to which BRGs satisfy some monotonicity properties, i.e., whether promoting some desired piece of belief to a given set of agents is actually always useful for making it accepted by all of them. We formally capture such a concept of promotion by a new family of belief change operators. We show that some basic monotonicity properties are not satisfied by BRGs in general, even when the agent's merging-based revision policies are fully rational (in the AGM sense). We also identify some classes where they hold.
【Keywords】:
【Paper Link】 【Pages】:1244-1250
【Authors】: Zohreh Shams ; Marina De Vos ; Nir Oren ; Julian Padget
【Abstract】: In a normative environment an agent's actions are not only directed by its goals but also by norms. Here, potential conflicts among the agent's goals and norms makes decision-making challenging. We therefore seek to answer the following questions: (i) how should an agent act in a normative environment? and (ii) how can the agent explain why it acted in a certain way? We propose a solution in which a normative planning problem serves as the basis for a practical reasoning approach based on argumentation. The properties of the best plan(s) w.r.t. goal achievement and norm compliance are mapped to arguments that are used to explain why a plan is justified, using a dialogue game.
【Keywords】:
【Paper Link】 【Pages】:1251-1257
【Authors】: Kostyantyn M. Shchekotykhin ; Thomas Schmitz ; Dietmar Jannach
【Abstract】: Model-Based Diagnosis is a principled approach to identify the possible causes when a system under observation behaves unexpectedly. In case the number of possible explanations for the unexpected behavior is large, sequential diagnosis approaches can be applied. The strategy of such approaches is to iteratively take additional measurements to narrow down the set of alternatives in order to find the true cause of the problem. In this paper we propose a sound and complete sequential diagnosis approach which does not require any information about the structure of the diagnosed system. The method is based on the new concept of "partial" diagnoses, which can be efficiently computed given a small number of minimal conflicts. As a result, the overall time needed for determining the best next measurement point can be significantly reduced. An experimental evaluation on different benchmark problems shows that our sequential diagnosis approach needs considerably less computation time when compared with an existing domain-independent approach.
【Keywords】:
【Paper Link】 【Pages】:1258-1264
【Authors】: Jason St. Jacques ; David Toman ; Grant E. Weddell
【Abstract】: We consider how SQL-like query languages over object-relational schemata canbe preserved in the setting of ontology based data access (OBDA), thus leveraging wide familiarity with relational technology. This is enabled by the adoption of the logic CFDnc-forall, a member of the CFD family of description logics (DLs). Of particular note is that this logic can fully simulate DLlite-F, a member ofthe DL-Lite family commonly used in the OBDA setting. Our main results present efficient algorithms that allow computation ofcertain answers with respect to CFDnc-forall nowledge bases, facilitating direct access to a pre-existing row-basedrelational encoding of the data without any need for mappings to triple-based representations.
【Keywords】:
【Paper Link】 【Pages】:1265-1271
【Authors】: Xingyu Su ; Marina Zanella ; Alban Grastien
【Abstract】: Diagnosability is the property that a Discrete-Event System (DES) exhibits if every fault can be detected and isolated within a finite number of (observable) events that have taken place after its occurrence. In the literature, diagnosability of DESs relies on the availability of a certain observation, which equals the sequence of observable events that have taken place in the DES. But can diagnosability be achieved even if the observation is uncertain? The present paper provides an answer to this question when the observation is temporally or logically uncertain, that is, when the order of the observed events or their (discrete) values are partially unknown. The original notion of compound observable event enables a smooth extension of both the definition of DES diagnosability in the literature and the twin plant method to check such a property. The intuition is to deal with a compound observable event the same way as with a single event. In case a DES is diagnosable even if its observation is uncertain, the diagnosis task can be performed (without any loss in the ability to identify every fault) although the available measuring equipment cannot get a certain observation.
【Keywords】:
【Paper Link】 【Pages】:1272-1278
【Authors】: Michael Thielscher
【Abstract】: Model sampling has proved to be a practically viable method for decision-making under uncertainty, for example in imperfect-information games with large state spaces. In this paper, we examine the logical foundations of sampling-based belief revision. We show that it satisfies six of the standard AGM postulates but not Vacuity nor Subexpansion. We provide a corresponding representation theorem that generalises the standard result from a single to a family of faithful assignments for a given belief set. We also provide a formal axiomatisation of sampling-based belief revision in the Situation Calculus as an alternative way of reasoning about actions, sensing, and beliefs.
【Keywords】:
【Paper Link】 【Pages】:1279-1285
【Authors】: Eleni Tsalapati ; Giorgos Stoilos ; Giorgos B. Stamou ; George Koletsos
【Abstract】: Inconsistent-tolerant semantics, like the IAR and ICAR semantics, have been proposed as means to compute meaningful query answers over inconsistent Description Logic (DL) ontologies. In the current paper we present a framework for scalable query answering under both the IAR and ICAR semantics, which is based on highly efficient data saturation systems. Our approach is sound and complete for ontologies expressed in the lightweight DL DL-Lite, but for more expressive DLs the problem is known to be intractable, hence our algorithm only computes upper approximations. Nevertheless, its structure motivates a new type of ICAR-like semantics which can be computed in polynomial time for a very large family of DLs. We have implemented our techniques and conducted an experimental evaluation obtaining encouraging results as both our IAR- and ICAR-answering approaches are far more efficient thanexisting available IAR-based answering systems.
【Keywords】:
【Paper Link】 【Pages】:1286-1292
【Authors】: Pieter Van Hertum ; Marcos Cramer ; Bart Bogaerts ; Marc Denecker
【Abstract】: In this paper we define and study an extension of autoepistemic logic (AEL) called distributed autoepistemic logic (dAEL) with multiple agents that have full introspection in their own knowledge as well as in that of others. This mutual full introspection between agents is motivated by an application of dAEL in access control. We define 2- and 3-valued semantic operators for dAEL. Using these operators, approximation fixpoint theory, an abstract algebraic framework that unifies different knowledge representation formalisms, immediately yields us a family of semantics for dAEL, each based on different intuitions that are well-studied in the context of AEL. The application in access control also motivates an extension of dAEL with inductive definitions (dAEL(ID)). We explain a use-case from access control to demonstrate how dAEL(ID) can be fruitfully applied to this domain and discuss how well-suited the different semantics are for the application in access control.
【Keywords】:
【Paper Link】 【Pages】:1293-1299
【Authors】: Zhigang Wang ; Juan-Zi Li
【Abstract】: Learning the representations of a knowledge graph has attracted significant research interest in the field of intelligent Web. By regarding each relation as one translation from head entity to tail entity, translation-based methods including TransE, TransH and TransR are simple, effective and achieving the state-of-the-art performance. However, they still suffer the following issues: (i) low performance when modeling 1-to-N, N-to-1 and N-to-N relations. (ii) limited performance due to the structure sparseness of the knowledge graph. In this paper, we propose a novel knowledge graph representation learning method by taking advantage of the rich context information in a text corpus. The rich textual context information is incorporated to expand the semantic structure of the knowledge graph and each relation is enabled to own different representations for different head and tail entities to better handle 1-to-N, N-to-1 and N-to-N relations. Experiments on multiple benchmark datasets show that our proposed method successfully addresses the above issues and significantly outperforms the state-of-the-art methods.
【Keywords】:
【Paper Link】 【Pages】:1300-1307
【Authors】: Jianfeng Wen ; Jianxin Li ; Yongyi Mao ; Shini Chen ; Richong Zhang
【Abstract】: The models developed to date for knowledge base embedding are all based on the assumption that the relations contained in knowledge bases are binary.For the training and testing of these embedding models, multi-fold (or n-ary) relational data are converted to triples (e.g., in FB15K dataset)and interpreted as instances of binary relations.This paper presents a canonical representation of knowledge bases containing multi-fold relations.We show that the existing embedding models on the popular FB15K datasets correspond to a suboptimal modelling framework, resulting in a loss of structural information. We advocate a novel modelling framework, which models multi-fold relations directly using this canonical representation.Using this framework, the existing TransH model is generalized to a new model, m-TransH. We demonstrate experimentally that m-TransH outperforms TransH by a large margin, thereby establishing a new state of the art.
【Keywords】:
【Paper Link】 【Pages】:1308-1314
【Authors】: Diedrich Wolter ; Jae Hee Lee
【Abstract】: This paper establishes new relationships between existing qualitative spatial and temporal representations. Qualitative spatial and temporal representation (QSTR) is concerned with abstractions of infinite spatial and temporal domains, which represent configurations of objects using a finite vocabulary of relations, also called a qualitative calculus. Classically, reasoning in QSTR is based on constraints. An important task is to identify decision procedures that are able to handle constraints from a single calculus or from several calculi. In particular the latter aspect is a longstanding challenge due to the multitude of calculi proposed. In this paper we consider propositional closures of qualitative constraints which enable progress with respect to the longstanding challenge. Propositional closure allows one to establish several translations between distinct calculi. This enables joint reasoning and provides new insights into computational complexity of individual calculi. We conclude that the study of propositional languages instead of previously considered purely relational languages is a viable research direction for QSTR leading to expressive formalisms and practical algorithms.
【Keywords】:
【Paper Link】 【Pages】:1315-1321
【Authors】: Han Xiao ; Minlie Huang ; Xiaoyan Zhu
【Abstract】: Knowledge graph embedding aims at offering a numerical knowledge representation paradigm by transforming the entities and relations into continuous vector space. However, existing methods could not characterize the knowledge graph in a fine degree to make a precise link prediction. There are two reasons: being an ill-posed algebraic system and adopting an overstrict geometric form. As precise link prediction is critical, we propose a manifold-based embedding principle (ManifoldE) which could be treated as a well-posed algebraic system that expands the position of golden triples from one point in current models to a manifold in ours. Extensive experiments show that the proposed models achieve substantial improvements against the state-of-the-art baselines especially for the precise prediction task, and yet maintain high efficiency.
【Keywords】:
【Paper Link】 【Pages】:1322-1329
【Authors】: Liping Xiong ; Yongmei Liu
【Abstract】: Strategy representation and reasoning for incomplete information concurrent games has recently received much attention in multi-agent system and AI communities. However, most of the logical frameworks are based on concrete game models, lack the abilities to reason about strategies explicitly or specify strategies procedurally, and ignore the issue of coordination within a coalition. In this paper, by a simple extension of a variant of multi-agent epistemic situation calculus with a strategy sort, we develop a general framework for strategy representation and reasoning for incomplete information concurrent games. Based on Golog, we propose a strategy programming language which can be conveniently used to specify collective strategies of coalitions at different granularities. We present a formalization of joint abilities of coalitions under commitments to strategy programs. Different kinds of individual strategic abilities can be distinguished in our framework. Both strategic abilities in ATL and joint abilities of Ghaderi et al. can be considered as joint abilities under special programs in our framework. We illustrate our work with a variant of Levesque's Squirrels World.
【Keywords】:
【Paper Link】 【Pages】:1330-1337
【Authors】: Heng Zhang ; Yan Zhang ; Jia-Huai You
【Abstract】: Existential rules, also known as data dependencies in Databases, have been recently rediscovered as a promising family of languages for Ontology-based Query Answering. In this paper, we prove that disjunctive embedded dependencies exactly capture the class of recursively enumerable ontologies in Ontology-based Conjunctive Query Answering (OCQA). Our expressive completeness result does not rely on any built-in linear order on the database. To establish the expressive completeness, we introduce a novel semantic definition for OCQA ontologies. We also show that neither the class of disjunctive tuple-generating dependencies nor the class of embedded dependencies is expressively complete for recursively enumerable OCQA ontologies.
【Keywords】:
【Paper Link】 【Pages】:1338-1344
【Authors】: Yuanlin Zhang ; Maede Rayatidamavandi
【Abstract】: The aggregates have greatly extended the representation power, in both theory and practice, of Answer Set Programming. Significant understanding of programs with aggregates has been gained in the last decade. However, there is still a substantial difficulty in understanding the semantics due to the nonmonotonic behavior of aggregates, which is demonstrated by several distinct semantics for aggregates in the existing work. In this paper, we aim to understand these distinct semantics in a more uniform way. Particularly, by satisfiability, rationality and consistency principles, we are able to give a uniform and simple characterizations of the three major distinct types of answer set semantics.
【Keywords】:
【Paper Link】 【Pages】:1345-1353
【Authors】: Yizheng Zhao ; Renate A. Schmidt
【Abstract】: Forgetting is a non-standard reasoning problem concerned with creating restricted views for ontologies relative to subsets of their initial signatures while preserving all logical consequences up to the symbols in the restricted views. In this paper, we present an Ackermann-based approach for forgetting of concept and role symbols in ontologies expressible in the description logic ALCOIHmu+(top,and). The method is one of only few approaches that can eliminate role symbols, that can handle role inverse, ABox statements, and is the only approach so far providing support for forgetting in description logics with nominals. Despite the inherent difficulty of forgetting for this level of expressivity, performance results with a prototypical implementation have shown very good success rates on real-world ontologies.
【Keywords】:
【Paper Link】 【Pages】:1354-1360
【Authors】: Irman Abdic ; Lex Fridman ; Daniel McDuff ; Erik Marchi ; Bryan Reimer ; Björn W. Schuller
【Abstract】: We present a method for detecting driver frustration from both video and audio streams captured during the driver's interaction with an in-vehicle voice-based navigation system. The video is of the driver's face when the machine is speaking, and the audio is of the driver's voice when he or she is speaking. We analyze a dataset of 20 drivers that contains 596 audio epochs (audio clips, with duration from 1 sec to 15 sec) and 615 video epochs (video clips, with duration from 1 sec to 45 sec). The dataset is balanced across 2 age groups, 2 vehicle systems, and both genders. The model was subject-independently trained and tested using 4-fold cross-validation. We achieve an accuracy of 77.4% for detecting frustration from a single audio epoch and 81.2% for detecting frustration from a single video epoch. We then treat the video and audio epochs as a sequence of interactions and use decision fusion to characterize the trade-off between decision time and classification accuracy, which improved the prediction accuracy to 88.5% after 9 epochs.
【Keywords】:
【Paper Link】 【Pages】:1361-1367
【Authors】: Eisa Alanazi ; Malek Mouhoub ; Sandra Zilles
【Abstract】: Learning of user preferences has become a core issue in AI research. For example, recent studies investigate learning of Conditional Preference Networks (CP-nets) from partial information. To assess the optimality of learning algorithms as well as to better understand the combinatorial structure of CP-net classes, it is helpful to calculate certain learning-theoretic information complexity parameters. This paper provides theoretical justification for exact values (or in some cases bounds) of some of the most central information complexity parameters, namely the VC dimension, the (recursive) teaching dimension, the self-directed learning complexity, and the optimal mistake bound, for classes of acyclic CP-nets. We further provide an algorithm that learns tree-structured CP-nets from membership queries. Using our results on complexity parameters, we can assess the optimality of our algorithm as well as that of another query learning algorithm for acyclic CP-nets presented in the literature.
【Keywords】:
【Paper Link】 【Pages】:1368-1374
【Authors】: Cesare Alippi ; Giacomo Boracchi ; Diego Carrera ; Manuel Roveri
【Abstract】: We address the problem of detecting changes in multivariate datastreams, and we investigate the intrinsic difficulty that change-detection methods have to face when the data dimension scales. In particular, we consider a general approach where changes are detected by comparing the distribution of the log-likelihood of the datastream over different time windows. Despite the fact that this approach constitutes the frame of several change-detection methods, its effectiveness when data dimension scales has never been investigated, which is indeed the goal of our paper. We show that the magnitude of the change can be naturally measured by the symmetric Kullback-Leibler divergence between the pre- and post-change distributions, and that the detectability of a change of a given magnitude worsens when the data dimension increases. This problem, which we refer to as detectability loss, is due to the linear relationship between the variance of the log-likelihood and the data dimension. We analytically derive the detectability loss on Gaussian-distributed datastreams, and empirically demonstrate that this problem holds also on real-world datasets and that can be harmful even at low data-dimensions (say, 10).
【Keywords】:
【Paper Link】 【Pages】:1375-1381
【Authors】: Ehsan Mohammady Ardehaly ; Aron Culotta ; Vivek Sundararaman ; Alwar Narayanan
【Abstract】: We investigate a suite of recommendation algorithms for audio news listening applications. This domain presents several challenges that distinguish it from more commonly studied applications such as movie recommendations: (1) we do not receive explicit rating feedback, instead only observing when a user skips a story; (2) new stories arrive continuously, increasing the importance of making recommendations for items with few observations (the cold start problem); (3) story attributes have high dimensionality, making it challenging to identify similar stories. To address the first challenge, we formulate the problem as predicting the percentage of a story a user will listen to; to address the remaining challenges, we propose several matrix factorization algorithms that cluster users, n-grams, and stories simultaneously, while optimizing prediction accuracy. We empirically evaluate our approach on a dataset of 50K users, 26K stories, and 975K interactions collected over a five month period. We find that while simple models work well for stories with many observations, our proposed approach performs best for stories with few ratings, which is critical for the real-world deployment of such an application.
【Keywords】:
【Paper Link】 【Pages】:1382-1388
【Authors】: Chao Chen ; Dongsheng Li ; Qin Lv ; Junchi Yan ; Stephen M. Chu ; Li Shang
【Abstract】: Matrix approximation (MA) is one of the most popular techniques for collaborative filtering (CF). Most existing MA methods train user/item latent factors based on a user-item rating matrix and then use the global latent factors to model all users/items. However, globally optimized latent factors may not reflect the unique interests shared among only subsets of users/items, without which unique interests of users may not be accurately modelled. As a result, existing MA methods, which cannot capture the uniqueness of different user/item, cannot provide optimal recommendation. In this paper, a mixture probabilistic matrix approximation (MPMA) method is proposed, which unifies globally optimized user/item feature vectors (on the entire rating matrix) and locally optimized user/item feature vectors (on subsets of user/item ratings) to improve recommendation accuracy. More specifically, in MPMA, a method is developed to find both globally and locally optimized user/item feature vectors. Then, a Gaussian mixture model is adopted to combine global predictions and local predictions to produce accurate rating predictions. Experimental study using MovieLens and Netflix datasets demonstrates that MPMA outperforms five state-of-the-art MA based CF methods in recommendation accuracy with good scalability.
【Keywords】:
【Paper Link】 【Pages】:1389-1395
【Authors】: Feng Chen ; Baojian Zhou
【Abstract】: Sparsity-constrained optimization is an important and challenging problem that has wide applicability in data mining, machine learning, and statistics. In this paper, we focus on sparsity-constrained optimization in cases where the cost function is a general nonlinear function and, in particular, the sparsity constraint is defined by a graph-structured sparsity model. Existing methods explore this problem in the context of sparse estimation in linear models. To the best of our knowledge, this is the first work to present an efficient approximation algorithm, namely, GRAPH-structured Matching Pursuit (GRAPH-MP), to optimize a general nonlinear function subject to graph-structured constraints. We prove that our algorithm enjoys the strong guarantees analogous to those designed for linear models in terms of convergence rate and approximation accuracy. As a case study, we specialize GRAPH-MP to optimize a number of well-known graph scan statistic models for the connected subgraph detection task, and empirical evidence demonstrates that our general algorithm performs superior over state-of-the-art methods that are designed specifically for the task of connected subgraph detection.
【Keywords】:
【Paper Link】 【Pages】:1396-1403
【Authors】: Ting Chen ; Lu An Tang ; Yizhou Sun ; Zhengzhang Chen ; Kai Zhang
【Abstract】: Anomaly detection plays an important role in modern data-driven security applications, such as detecting suspicious access to a socket from a process. In many cases, such events can be described as a collection of categorical values that are considered as entities of different types, which we call heterogeneous categorical events. Due to the lack of intrinsic distance measures among entities, and the exponentially large event space, most existing work relies heavily on heuristics to calculate abnormal scores for events. Different from previous work, we propose a principled and unified probabilistic model APE (Anomaly detection via Probabilistic pairwise interaction and Entity embedding) that directly models the likelihood of events. In this model, we embed entities into a common latent space using their observed co-occurrence in different events. More specifically, we first model the compatibility of each pair of entities according to their embeddings. Then we utilize the weighted pairwise interactions of different entity types to define the event probability. Using Noise-Contrastive Estimation with "context-dependent" noise distribution, our model can be learned efficiently regardless of the large event space. Experimental results on real enterprise surveillance data show that our methods can accurately detect abnormal events compared to other state-of-the-art abnormal detection techniques.
【Keywords】:
【Paper Link】 【Pages】:1404-1410
【Authors】: Yu Chen ; Tom Diethe ; Peter A. Flach
【Abstract】: We present an unsupervised approach for discovery of Activities of Daily Living (ADL) in a smart home. Activity discovery is an important enabling technology, for example to tackle the healthcare requirements of elderly people in their homes. The technique applied most often is supervised learning, which relies on expensive labelled data and lacks the flexibility to discover unseen activities. Building on ideas from text mining, we present a powerful topic model and a segmentation algorithm that can learn from unlabelled sensor data. The model has been evaluated extensively on datasets collected from real smart homes. The results demonstrate that this approach can successfully discover the activities of residents, and can be effectively used in a range of applications such as detection of abnormal activities and monitoring of sleep quality, among many others.
【Keywords】:
【Paper Link】 【Pages】:1411-1417
【Authors】: Yu-An Chung ; Hsuan-Tien Lin ; Shao-Wen Yang
【Abstract】: Deep learning has been one of the most prominent machine learning techniques nowadays, being the state-of-the-art on a broad range of applications where automatic feature extraction is needed. Many such applications also demand varying costs for different types of mis-classification errors, but it is not clear whether or how such cost information can be incorporated into deep learning to improve performance. In this work, we first design a novel loss function that embeds the cost information for the training stage of cost-sensitive deep learning. We then show that the loss function can also be integrated into the pre-training stage to conduct cost-aware feature extraction more effectively. Extensive experimental results justify the validity of the novel loss function for making existing deep learning models cost-sensitive, and demonstrate that our proposed model with cost-aware pre-training and training outperforms non-deep models and other deep models that digest the cost information in other stages.
【Keywords】:
【Paper Link】 【Pages】:1418-1424
【Authors】: Andrew Cropper ; Stephen H. Muggleton
【Abstract】: Many tasks in AI require the design of complex programs and representations, whether for programming robots, designing game-playing programs, or conducting textual or visual transformations. This paper explores a novel inductive logic programming approach to learn such programs from examples. To reduce the complexity of the learned programs, and thus the search for such a program, we introduce higher-order operations involving an alternation of Abstraction and Invention. Abstractions are described using logic program definitions containing higher-order predicate variables. Inventions involve the construction of definitions for the predicate variables used in the Abstractions. The use of Abstractions extends the Meta-Interpretive Learning framework and is supported by the use of a user-extendable set of higher-order operators, such as map, until, and ifthenelse. Using these operators reduces the textual complexity required to express target classes of programs. We provide sample complexity results which indicate that the approach leads to reductions in the numbers of examples required to reach high predictive accuracy, as well as significant reductions in overall learning time. Our experiments demonstrate increased accuracy and reduced learning times in all cases. We believe that this paper is the first in the literature to demonstrate the efficiency and accuracy advantages involved in the use of higher-order abstractions.
【Keywords】:
【Paper Link】 【Pages】:1425-1431
【Authors】: Gary Doran ; Andrew Latham ; Soumya Ray
【Abstract】: We study the problem of bag-level classification from generalized multiple-instance (GMI) data. GMI learning is an extension of the popular multiple-instance setting. In GMI data, bags are labeled positive if they contain instances of certain types, and avoid instances of other types. For example, an image of a "sunny beach"' should contain sand and sea, but not clouds. We formulate a novel generative process for the GMI setting in which bags are distributions over instances. In this model, we show that a broad class of distribution-distance kernels is sufficient to represent arbitrary GMI concepts. Further, we show that a variety of previously proposed kernel approaches to the standard MI and GMI settings can be unified under the distribution kernel framework. We perform an extensive empirical study which indicates that the family of distribution distance kernels is accurate for a wide variety of real-world MI and GMI tasks as well as efficient when compared to a large set of baselines. Our theoretical and empirical results indicate that distribution-distance kernels can serve as a unifying framework for learning bag labels from GMI (and therefore MI) problems.
【Keywords】:
【Paper Link】 【Pages】:1432-1440
【Authors】: Finale Doshi-Velez ; George Konidaris
【Abstract】: Control applications often feature tasks with similar, but not identical, dynamics. We introduce the Hidden Parameter Markov Decision Process (HiP-MDP), a framework that parametrizes a family of related dynamical systems with a low-dimensional set of latent factors, and introduce a semiparametric regression approach for learning its structure from data. We show that a learned HiP-MDP rapidly identifies the dynamics of new task instances in several settings, flexibly adapting to task variation.
【Keywords】:
【Paper Link】 【Pages】:1441-1447
【Authors】: Ahmed Elbagoury ; Rania Ibrahim ; Mohamed S. Kamel ; Fakhri Karray
【Abstract】: With the rapid increase in the available data, it becomes computationally harder to extract useful information. Thus, several techniques like PCA were proposed to embed high-dimensional data into low-dimensional latent space. However, these techniques don't take the data relations into account. This motivated the development of other techniques like MDS and LLE which preserve the relations between the data instances. Nonetheless, all these techniques still use latent features, which are difficult for data analysts to understand and grasp the information encoded in them. In this work, a new embedding technique is proposed to mitigate the previous problems by projecting the data to a space described by few points (i.e, exemplars) which preserves the relations between the data points. The proposed method Exemplar-based Kernel Preserving (EBEK) embedding is shown theoretically to achieve the lowest reconstruction error of the kernel matrix. Using EBEK in approximate nearest neighbor task shows its ability to outperform related work by up to 60% in the recall while maintaining a good running time. In addition, our interpretability experiments show that EBEK's selected basis are more understandable than the latent basis in images datasets.
【Keywords】:
【Paper Link】 【Pages】:1448-1454
【Authors】: Karen Braga Enes ; Saulo Moraes Villela ; Raul Fonseca Neto
【Abstract】: A classifier able to minimize the generalization error of a particular problem for any set of unseen samples is named Bayes-optimal classifier. The hypothesis induced by such classifier is equivalent to the optimal Bayes point, which is approximately equivalent to the center of mass of the version space. However, there are only a few methods for estimating the center of mass and most of them are computationally expensive or impractical, especially for large datasets. In this paper we present the Version Space Reduction Machine (VSRM), a new method that obtains an approximation of the center of mass. The method works by means of successive reductions of the version space which are consistent with an oracle's decision. This oracle is represented by the majority voting of an ensemble, whose components must contain a reasonable diversity level to ensure an effective approximation. We conduct an experimental study on microarray datasets and assess the performance of the proposed method compared to Support Vector Machine and Bayes Point Machine. Our method consistently outperforms the others. Such result indicates that the proposed method provides a better approximation of the center of mass.
【Keywords】:
【Paper Link】 【Pages】:1455-1461
【Authors】: Sarah M. Erfani ; Mahsa Baktashmotlagh ; Masud Moshtaghi ; Vinh Nguyen ; Christopher Leckie ; James Bailey ; Kotagiri Ramamohanarao
【Abstract】: Many conventional statistical machine learning algorithms generalise poorly if distribution bias exists in the datasets. For example, distribution bias arises in the context of domain generalization, where knowledge acquired from multiple source domains need to be used in a previously unseen target domains. We propose Elliptical Summary Randomisation (ESRand), an efficient domain generalisation approach that comprises of a randomized kernel and elliptical data summarisation. ESRand learns a domain interdependent projection to a latent subspace that minimises the existing biases to the data while maintaining the functional relationship between domains. In the latent subspace, ellipsoidal summaries replace the samples to enhance the generalisation by further removing biasand noise in the data. Moreover, the summarization enables large-scale data processing by significantly reducing the size of the data. Through comprehensive analysis, we show that our subspace-based approach outperforms state-of-the-art results on several activity recognition benchmark datasets, while keeping the computational complexity significantly low.
【Keywords】:
【Paper Link】 【Pages】:1462-1468
【Authors】: Xuhui Fan ; Richard Yi Da Xu ; Longbing Cao
【Abstract】: The Mixed-Membership Stochastic Blockmodels (MMSB) is a popular framework for modelling social relationships by fully exploiting each individual node's participation (or membership) in a social network. Despite its powerful representations, MMSB assumes that the membership indicators of each pair of nodes (i.e., people) are distributed independently. However, such an assumption often does not hold in real-life social networks, in which certain known groups of people may correlate with each other in terms of factors such as their membership categories. To expand MMSB's ability to model such dependent relationships, a new framework - a Copula Mixed-Membership Stochastic Blockmodel - is introduced in this paper for modeling intra-group correlations, namely an individual Copula function jointly models the membership pairs of those nodes within the group of interest. This framework enables various Copula functions to be used on demand, while maintaining the membership indicator's marginal distribution needed for modelling membership indicators with other nodes outside of the group of interest. Sampling algorithms for both the finite and infinite number of groups are also detailed. Our experimental results show its superior performance in capturing group interactions when compared with the baseline models on both synthetic and real world datasets.
【Keywords】:
【Paper Link】 【Pages】:1469-1475
【Authors】: Jie Fu ; Hongyin Luo ; Jiashi Feng ; Kian Hsiang Low ; Tat-Seng Chua
【Abstract】: The performance of deep neural networks is well-known to be sensitive to the setting of their hyperparameters. Recent advances in reverse-mode automatic differentiation allow for optimizing hyperparameters with gradients. The standard way of computing these gradients involves a forward and backward pass of computations. However, the backward pass usually needs to consume unaffordable memory to store all the intermediate variables to exactly reverse the forward training procedure. In this work we propose a simple but effective method, DrMAD, to distill the knowledge of the forward pass into a shortcut path, through which we approximately reverse the training trajectory. Experiments on several image benchmark datasets show that DrMAD is at least 45 times faster and consumes 100 times less memory compared to state-of-the-art methods for optimizing hyperparameters with minimal compromise to its effectiveness. To the best of our knowledge, DrMAD is the first research attempt to make it practical to automatically tune thousands of hyperparameters of deep neural networks. The code can be downloaded from https://github.com/bigaidream-projects/drmad
【Keywords】:
【Paper Link】 【Pages】:1476-1482
【Authors】: Junning Gao ; Makoto Yamada ; Samuel Kaski ; Hiroshi Mamitsuka ; Shanfeng Zhu
【Abstract】: We formulate ensemble clustering as a regularization problem over nuclear norm and cluster-wise group norm, and present an efficient optimization algorithm, which we call Robust Convex Ensemble Clustering (RCEC). A key feature of RCEC allows to remove anomalous cluster assignments obtained from component clustering methods by using the group-norm regularization. Moreover, the proposed method is convex and can find the globally optimal solution. We first showed that using synthetic data experiments, RCEC could learn stable cluster assignments from the input matrix including anomalous clusters. We then showed that RCEC outperformed state-of-the-art ensemble clustering methods by using real-world data sets.
【Keywords】:
【Paper Link】 【Pages】:1483-1489
【Authors】: Li Gao ; Jia Wu ; Hong Yang ; Zhi Qiao ; Chuan Zhou ; Yue Hu
【Abstract】: Network coarsening refers to a new class of graph `zoom-out' operations by grouping similar nodes and edges together so that a smaller equivalent representation of the graph can be obtained for big network analysis. Existing network coarsening methods consider that network structures are static and thus cannot handle dynamic networks. On the other hand, data-driven approaches can infer dynamic network structures by using network information spreading data. However, existing data-driven approaches neglect static network structures that are potentially useful for inferring big networks. In this paper, we present a new semi-data-driven network coarsening model to learn coarsened networks by embedding both static network structure data and dynamic network information spreading data. We prove that the learning model is convex and the Accelerated Proximal Gradient algorithm is adapted to achieve the global optima. Experiments on both synthetic and real-world data sets demonstrate the quality and effectiveness of the proposed method.
【Keywords】:
【Paper Link】 【Pages】:1490-1496
【Authors】: Tian Gao ; Qiang Ji
【Abstract】: For many applications, the observed data may be incomplete and there often exist variables that are unobserved but play an important role in capturing the underlying relationships. In this work, we propose a method to identify local latent variables and to determine their structural relations with the observed variables. We formulate the local latent variable discovery as discovering the Markov Blanket (MB) of a target variable. To efficiently search the latent variable space, we exploit MB topology to divide the latent space into different subspaces. Within each subspace, we employ a constrained structure expectation-maximization algorithm to greedily learn the MB with latent variables. We evaluate the performance of our method on synthetic data to demonstrate its effectiveness in identifying the correct latent variables. We further apply our algorithm to feature discovery and selection problem, and show that the latent variables learned through the proposed method can improve the classification accuracy in benchmark feature selection and discovery datasets.
【Keywords】:
【Paper Link】 【Pages】:1497-1504
【Authors】: Martin Gebser ; Thomas Guyet ; René Quiniou ; Javier Romero ; Torsten Schaub
【Abstract】: We introduce a framework for knowledge-based sequence mining, based on Answer Set Programming (ASP). We begin by modeling the basic task and refine it in the sequel in several ways. First, we show how easily condensed patterns can be extracted by modular extensions of the basic approach. Second, we illustrate how ASP's preference handling capacities can be exploited for mining patterns of interest. In doing so, we demonstrate the ease of incorporating knowledge into the ASP-based mining process. To assess the trade-off in effectiveness, we provide an empirical study comparing our approach with a related sequence mining mechanism.
【Keywords】:
【Paper Link】 【Pages】:1505-1511
【Authors】: Clement Gehring ; Yangchen Pan ; Martha White
【Abstract】: Balancing between computational efficiency and sample efficiency is an important goal in reinforcement learning. Temporal difference (TD) learning algorithms stochastically update the value function, with a linear time complexity in the number of features, whereas least-squares temporal difference (LSTD) algorithms are sample efficient but can be quadratic in the number of features. In this work, we develop an efficient incremental low-rank LSTD(λ) algorithm that progresses towards the goal of better balancing computation and sample efficiency. The algorithm reduces the computation and storage complexity to the number of features times the chosen rank parameter while summarizing past samples efficiently to nearly obtain the sample efficiency of LSTD. We derive a simulation bound on the solution given by truncated low-rank approximation, illustrating a bias-variance trade-off dependent on the choice of rank. We demonstrate that the algorithm effectively balances computational complexity and sample efficiency for policy evaluation in a benchmark task and a high-dimensional energy allocation domain.
【Keywords】:
【Paper Link】 【Pages】:1512-1518
【Authors】: Hongliang Guo ; Jie Zhang
【Abstract】: With the rapid development of data sensing and collection technologies, we can easily obtain large volumes of data (big data). However, big data poses huge challenges to many popular machine learning techniques which take all the data at the same time for processing. To address the big data related challenges, we first partition the data along its feature space, and apply the parallel block coordinate descent algorithm for distributed computation; then, we continue to partition the data along the sample space, and propose a novel matrix decomposition and combination approach for distributed processing. The final results from all the entities are guaranteed to be the same as the centralized solution. Extensive experiments performed on Hadoop confirm that our proposed approach is superior in terms of both testing errors and convergence rate (computation time) over the canonical distributed machine learning techniques that deal with big data.
【Keywords】:
【Paper Link】 【Pages】:1519-1525
【Authors】: Xiaoxiao Guo ; Satinder P. Singh ; Richard L. Lewis ; Honglak Lee
【Abstract】: Monte Carlo Tree Search (MCTS) methods have proven powerful in planning for sequential decision-making problems such as Go and video games, but their performance can be poor when the planning depth and sampling trajectories are limited or when the rewards are sparse. We present an adaptation of PGRD (policy-gradient for reward-design) for learning a reward-bonus function to improve UCT (a MCTS algorithm). Unlike previous applications of PGRD in which the space of reward-bonus functions was limited to linear functions of hand-coded state-action-features, we use PGRD with a multi-layer convolutional neural network to automatically learn features from raw perception as well as to adapt the non-linear reward-bonus function parameters. We also adopt a variance-reducing gradient method to improve PGRD's performance. The new method improves UCT's performance on multiple ATARI games compared to UCT without the reward bonus. Combining PGRD and Deep Learning in this way should make adapting rewards for MCTS algorithms far more widely and practically applicable than before.
【Keywords】:
【Paper Link】 【Pages】:1526-1532
【Authors】: Yuchen Guo ; Guiguang Ding ; Yue Gao ; Jianmin Wang
【Abstract】: To save the labeling efforts for training a classification model, we can simultaneously adopt Active Learning (AL) to select the most informative samples for human labeling, and Semi-supervised Learning (SSL) to construct effective classifiers using a few labeled samples and a large number of unlabeled samples. Recently, using Transfer Learning (TL) to enhance AL and SSL, i.e., T-SS-AL, has gained considerable attention. However, existing T-SS-AL methods mostly focus on the situation where the source domain and the target domain share the same classes. In this paper, we consider a more practical and challenging setting where the source domain and the target domain have different but related classes. We propose a novel cross-class sample transfer based T-SS-AL method, called CC-SS-AL, to exploit the information from the source domain. Our key idea is to select samples from the source domain which are very similar to the target domain classes and assign pseudo labels to them for classifier training. Extensive experiments on three datasets verify the efficacy of the proposed method.
【Keywords】:
【Paper Link】 【Pages】:1533-1540
【Authors】: Nils Y. Hammerla ; Shane Halloran ; Thomas Plötz
【Abstract】: Human activity recognition (HAR) in ubiquitous computing is beginning to adopt deep learning to substitute for well-established analysis techniques that rely on hand-crafted feature extraction and classification methods. However, from these isolated applications of custom deep architectures it is difficult to gain an overview of their suitability for problems ranging from the recognition of manipulative gestures to the segmentation and identification of physical activities like running or ascending stairs. In this paper we rigorously explore deep, convolutional, and recurrent approaches across three representative datasets that contain movement data captured with wearable sensors. We describe how to train recurrent approaches in this setting, introduce a novel regularisation approach, and illustrate how they outperform the state-of-the-art on a large benchmark dataset. We investigate the suitability of each model for HAR, across thousands of recognition experiments with randomly sampled model configurations, explore the impact of hyperparameters using the fANOVA framework, and provide guidelines for the practitioner who wants to apply deep learning in their problem setting.
【Keywords】:
【Paper Link】 【Pages】:1541-1547
【Authors】: Tao Han ; Hailong Sun ; Yangqiu Song ; Yili Fang ; Xudong Liu
【Abstract】: Crowdsourcing has been a helpful mechanism to leverage human intelligence to acquire useful knowledge for well defined tasks. However, when aggregating the crowd knowledge based on the currently developed voting algorithms, it often results in common knowledge that may not be expected. In this paper, we consider the problem of collecting as specific as possible knowledge via crowdsourcing. With the help of using external knowledge base such as WordNet, we incorporate the semantic relations between the alternative answers into a probabilistic model to determine which answer is more specific. We formulate the probabilistic model considering both worker's ability and task's difficulty, and solve it by expectation-maximization (EM) algorithm. Experimental results show that our approach achieved 35.88% improvement over majority voting when more specific answers are expected.
【Keywords】:
【Paper Link】 【Pages】:1548-1554
【Authors】: Yufei Han ; Yun Shen
【Abstract】: Selecting discriminative features in positive unlabelled (PU) learning tasks is a challenging problem due to lack of negative class information. Traditional supervised and semi-supervised feature selection methods are not able to be applied directly in this scenario, and unsupervised feature selection algorithms are designed to handle unlabelled data while neglecting the available information from positive class. To leverage the partially observed positive class information, we propose to encode the weakly supervised information in PU learning tasks into pairwise constraints between training instances. Violation of pairwise constraints are measured and incorporated into a partially supervised graph embedding model. Extensive experiments on different benchmark databases and a real-world cyber security application demonstrate the effectiveness of our algorithm.
【Keywords】:
【Paper Link】 【Pages】:1555-1561
【Authors】: Jia He ; Changying Du ; Fuzhen Zhuang ; Xin Yin ; Qing He ; Guoping Long
【Abstract】: Last decades have witnessed a number of studies devoted to multi-view learning algorithms, however, few efforts have been made to handle online multi-view learning scenarios. In this paper, we propose an online Bayesian multi-view learning algorithm to learn predictive subspace with max-margin principle. Specifically, we first define the latent margin loss for classification in the subspace, and then cast the learning problem into a variational Bayesian framework by exploiting the pseudo-likelihood and data augmentation idea. With the variational approximate posterior inferred from the past samples, we can naturally combine historical knowledge with new arrival data, in a Bayesian Passive-Aggressive style. Experiments on various classification tasks show that our model have superior performance.
【Keywords】:
【Paper Link】 【Pages】:1562-1570
【Authors】: Hoda Heidari ; Michael Kearns ; Aaron Roth
【Abstract】: We consider a variant of the well-studied multi-armed bandit problem in which the reward from each action evolves monotonically in the number of times the decision maker chooses to take that action. We are motivated by settings in which we must give a series of homogeneous tasks to a finite set of arms (workers) whose performance may improve (due to learning) or decay (due to loss of interest) with repeated trials. We assume that the arm-dependent rates at which the rewards change are unknown to the decision maker, and propose algorithms with provably optimal policy regret bounds, a much stronger notion than the often-studied external regret. For the case where the rewards are increasing and concave, we give an algorithm whose policy regret is sublinear and has a (provably necessary) dependence on the time required to distinguish the optimal arm from the rest. We illustrate the behavior and performance of this algorithm via simulations. For the decreasing case, we present a simple greedy approach and show that the policy regret of this algorithm is constant and upper bounded by the number of arms.
【Keywords】:
【Paper Link】 【Pages】:1571-1577
【Authors】: Teakgyu Hong ; Jongmin Lee ; Kee-Eung Kim ; Pedro A. Ortega ; Daniel D. Lee
【Abstract】: In the standard reinforcement learning setting, the agent learns optimal policy solely from state transitions and rewards from the environment. We consider an extended setting where a trainer additionally provides feedback on the actions executed by the agent. This requires appropriately incorporating the feedback, even when the feedback is not necessarily accurate. In this paper, we present a Bayesian approach to this extended reinforcement learning setting. Specifically, we extend Kalman Temporal Difference learning to compute the posterior distribution over Q-values given the state transitions and rewards from the environment as well as the feedback from the trainer. Through experiments on standard reinforcement learning tasks, we show that learning performance can be significantly improved even with inaccurate feedback.
【Keywords】:
【Paper Link】 【Pages】:1578-1584
【Authors】: Zhiting Hu ; Gang Luo ; Mrinmaya Sachan ; Eric P. Xing ; Zaiqing Nie
【Abstract】: Topic models represent latent topics as probability distributions over words which can be hard to interpret due to the lack of grounded semantics. In this paper, we propose a structured topic representation based on an entity taxonomy from a knowledge base. A probabilistic model is developed to infer both hidden topics and entities from text corpora. Each topic is equipped with a random walk over the entity hierarchy to extract semantically grounded and coherent themes. Accurate entity modeling is achieved by leveraging rich textual features from the knowledge base. Experiments show significant superiority of our approach in topic perplexity and key entity identification, indicating potentials of the grounded modeling for semantic extraction and language understanding applications.
【Keywords】:
【Paper Link】 【Pages】:1585-1591
【Authors】: Long-Kai Huang ; Sinno Jialin Pan
【Abstract】: Learning to hash has become a crucial technique for big data analytics. Among existing methods, supervised learning approaches play an important role as they can produce compact codes and enable semantic search. However, the size of an instance pairwise similarity matrix used in most supervised hashing methods is quadratic to the size of labeled training data, which is very expensive in terms of space, especially for a large-scale learning problem. This limitation hinders the full utilization of labeled instances for learning a more precise hashing model. To overcome this limitation, we propose a class-wise supervised hashing method that trains a model based on a class-pairwise similarity matrix, whose size is much smaller than the instance-pairwise similarity matrix in many applications. In addition, besides a set of hash functions, our proposed method learns a set of classwise code-prototypes with active bits for different classes. These class-wise code-prototypes can help to learn more precise compact codes for semantic information retrieval. Experimental results verify the superior effectiveness of our proposed method over other baseline hashing methods.
【Keywords】:
【Paper Link】 【Pages】:1592-1598
【Authors】: Sheng-Jun Huang ; Songcan Chen
【Abstract】: To learn with limited labeled data, active learning tries to query more labels from an oracle, while transfer learning tries to utilize the labeled data from a related source domain. However, in many real cases, there is very few labeled data in both source and target domains, and the oracle is unavailable in the target domain. To solve this practical yet rarely studied problem, in this paper, we jointly perform transfer learning and active learning by querying the most valuable information from the source domain. The computation of importance weights for domain adaptation and the instance selection for active queries are integrated into one unified framework based on distribution matching, which is further solved with alternating optimization. The effectiveness of the proposed method is validated by experiments on 15 datasets for sentiment analysis and text categorization.
【Keywords】:
【Paper Link】 【Pages】:1599-1605
【Authors】: Wen-bing Huang ; Le-le Cao ; Fuchun Sun ; Deli Zhao ; Huaping Liu ; Shanshan Yu
【Abstract】: Standard subspace algorithms learn Linear Dynamical Systems (LDSs) from time series with the least-square method, where the stability of the system is not naturally guaranteed. In this paper, we propose a novel approach for learning stable systems by enforcing stability directly on the least-square solutions. To this end, we first explore the spectral-radius property of the least-square transition matrix and then determine the key component that incurs the instability of the transition matrix. By multiplying the unstable component with a weight matrix on the right side, we obtain a weighted-least-square transition matrix that is further optimized to minimize the reconstruction error of the state sequence while still maintaining the stable constraint. Comparative experimental evaluations demonstrate that our proposed methods outperform the state-of-the-art methods regarding the reconstruction accuracy and the learning efficiency.
【Keywords】:
【Paper Link】 【Pages】:1606-1612
【Authors】: Xuan Huo ; Ming Li ; Zhi-Hua Zhou
【Abstract】: Bug reports provide an effective way for end-users to disclose potential bugs hidden in a software system, while automatically locating the potential buggy source code according to a bug report remains a great challenge in software maintenance. Many previous studies treated the source code as natural language by representing both the bug report and source code based on bag-of-words feature representations, and correlate the bug report and source code by measuring similarity in the same lexical feature space. However, these approaches fail to consider the structure information of source code which carries additional semantics beyond the lexical terms. Such information is important in modeling program functionality. In this paper, we propose a novel convolutional neural network NP-CNN, which leverages both lexical and program structure information to learn unified features from natural language and source code in programming language for automatically locating the potential buggy source code according to bug report. Experimental results on widely-used software projects indicate that NP-CNN significantly outperforms the state-of-the-art methods in locating the buggy source files.
【Keywords】:
【Paper Link】 【Pages】:1613-1619
【Authors】: Tsuyoshi Idé ; Dzung T. Phan ; Jayant Kalagnanam
【Abstract】: This paper addresses the task of change detection from noisy multivariate time-series data. One major feature of our approach is to leverage directional statistics as the noise-robust signature of time-series data. To capture major patterns, we introduce a regularized maximum likelihood equation for the von Mises-Fisher distribution, which simultaneously learns directional statistics and sample weights to filter out unwanted samples contaminated by the noise. We show that the optimization problem is reduced to the trust region subproblem in a certain limit, where global optimality is guaranteed. To evaluate the amount of changes, we introduce a novel distance measure on the Stiefel manifold. The method is validated with real-world data from an ore mining system.
【Keywords】:
【Paper Link】 【Pages】:1620-1626
【Authors】: David Isele ; Mohammad Rostami ; Eric Eaton
【Abstract】: Knowledge transfer between tasks can improve the performance of learned models, but requires an accurate estimate of the inter-task relationships to identify the relevant knowledge to transfer. These inter-task relationships are typically estimated based on training data for each task, which is inefficient in lifelong learning settings where the goal is to learn each consecutive task rapidly from as little data as possible. To reduce this burden, we develop a lifelong reinforcement learning method based on coupled dictionary learning that incorporates high-level task descriptors to model the inter-task relationships. We show that using task descriptors improves the performance of the learned task policies, providing both theoretical justification for the benefit and empirical demonstration of the improvement across a variety of dynamical control problems. Given only the descriptor for a new task, the lifelong learner is also able to accurately predict the task policy through zero-shot learning using the coupled dictionary, eliminating the need to pause to gather training data before addressing the task.
【Keywords】:
【Paper Link】 【Pages】:1627-1633
【Authors】: Ling Jian ; Jundong Li ; Kai Shu ; Huan Liu
【Abstract】: Multi-label learning has been extensively studied in the area of bioinformatics, information retrieval, multimedia annotation, etc. In multi-label learning, each instance is associated with multiple interdependent class labels, the label information can be noisy and incomplete. In addition, multi-labeled data often has noisy, irrelevant and redundant features of high dimensionality. As an effective data preprocessing step, feature selection has shown its effectiveness to prepare high-dimensional data for numerous data mining and machine learning tasks. Most of existing multi-label feature selection algorithms either boil down to solving multiple single-labeled feature selection problems or directly make use of imperfect labels. Therefore, they may not be able to find discriminative features that are shared by multiple labels. In this paper, we propose a novel multi-label informed feature selection framework MIFS, which exploits label correlations to select discriminative features across multiple labels. Specifically, to reduce the negative effects of imperfect label information in finding label correlations, we decompose the multi-label information into a low-dimensional space and then employ the reduced space to steer the feature selection process. Empirical studies on real-world datasets demonstrate the effectiveness and efficiency of the proposed framework.
【Keywords】:
【Paper Link】 【Pages】:1634-1639
【Authors】: Bo Jiang ; Chris H. Q. Ding ; Bin Luo
【Abstract】: Trace norm based rank regularization techniques have been successfully applied to learn a low-rank recovery for high-dimensional noise data. In many applications, it is desirable to add new samples to previously recovered data which is known as out of sample data recovery problem. However, traditional trace norm based regularization methods can not naturally cope with new samples and thus fail to deal with out-of-sample data recovery. In this paper, we propose a new robust out-of-sample data recovery (ROSR) model for trace norm based regularization methods. An effective iterative algorithm, with the proof of convergence, is presented to find the optimal solution of ROSR problem. As an application, we apply our ROSR to image classification task. Experimental results on six image datasets demonstrate the effectiveness and benefits of the proposed ROSR method.
【Keywords】:
【Paper Link】 【Pages】:1640-1647
【Authors】: Nan Jiang ; Satinder P. Singh ; Ambuj Tewari
【Abstract】: Planning in MDPs often uses a smaller planning horizon than specified in the problem to save computational expense at the risk of a loss due to suboptimal plans. Jiang et al. [2015b] recently showed that smaller than specified planning horizons can in fact be beneficial in cases where the MDP model is learned from data and therefore not accurate. In this paper, we consider planning with accurate models and investigate structural properties of MDPs that bound the loss incurred by using smaller than specified planning horizons. We identify a number of structural parameters some of which depend on the reward function alone, some on the transition dynamics alone, and some that depend on the interaction between rewards and transition dynamics. We provide planning loss bounds in terms of these structural parameters and, in some cases, also show tightness of the upper bounds. Empirical results with randomly generated MDPs are used to validate qualitative properties of our theoretical bounds for shallow planning.
【Keywords】:
【Paper Link】 【Pages】:1648-1654
【Authors】: George Konidaris
【Abstract】: We describe a framework for building abstraction hierarchies whereby an agent alternates skill- and representation-construction phases to construct a sequence of increasingly abstract Markov decision processes. Our formulation builds on recent results showing that the appropriate abstract representation of a problem is specified by the agent's skills. We describe how such a hierarchy can be used for fast planning, and illustrate the construction of an appropriate hierarchy for the Taxi domain.
【Keywords】:
【Paper Link】 【Pages】:1655-1661
【Authors】: Ondrej Kuzelka ; Yuyi Wang ; Jan Ramon
【Abstract】: This paper deals with the generalization ability of classifiers trained from non-iid evolutionary-related data in which all training and testing examples correspond to leaves of a phylogenetic tree. For the realizable case, we prove PAC-type upper and lower bounds based on symmetries and matchings in such trees.
【Keywords】:
【Paper Link】 【Pages】:1662-1668
【Authors】: Lucas Langer ; Borja Balle ; Doina Precup
【Abstract】: Recent years have seen the development of efficient and provably correct spectral algorithms for learning models of partially observable environments arising in many applications. But despite the high hopes raised by this new class of algorithms, their practical impact is still below expectations. One reason for this is the difficulty in adapting spectral methods to exploit structural constraints about different target environments which can be known beforehand. A natural structure intrinsic to many dynamical systems is a multi-resolution behaviour where interesting phenomena occur at different time scales during the evolution of the system. In this paper we introduce the multi-step predictive state representation (M-PSR) and an associated learning algorithm that finds and leverages frequent patterns of observations at multiple scales in dynamical systems with discrete observations. We perform experiments on robot exploration tasks in a wide variety of environments and conclude that the use of M-PSRs improves over the classical PSR for varying amounts of data, environment sizes, and number of observations symbols.
【Keywords】:
【Paper Link】 【Pages】:1669-1675
【Authors】: Sang-Woo Lee ; Chung-yeon Lee ; Dong-Hyun Kwak ; Jiwon Kim ; Jeonghee Kim ; Byoung-Tak Zhang
【Abstract】: Learning from human behaviors in the real world is important for building human-aware intelligent systems such as personalized digital assistants and autonomous humanoid robots.Everyday activities of human life can now be measured through wearable sensors.However, innovations are required to learn these sensory data in an online incremental manner over an extended period of time. Here we propose a dual memory architecture that processes slow-changing global patterns as well as keeps track of fast-changing local behaviors over a lifetime. The lifelong learnability is achieved by developing new techniques, such as weight transfer and an online learning algorithm with incremental features. The proposed model outperformed other comparable methods on two real-life data-sets: the image-stream dataset and the real-world lifelogs collected through the Google Glass for 46 days.
【Keywords】:
【Paper Link】 【Pages】:1676-1682
【Authors】: Yoad Lewenberg ; Yoram Bachrach ; Sukrit Shankar ; Antonio Criminisi
【Abstract】: We consider the task of predicting various traits of a person given an image of their face. We estimate both objective traits, such as gender, ethnicity and hair-color; as well as subjective traits, such as the emotion a person expresses or whether he is humorous or attractive. For sizeable experimentation, we contribute a new Face Attributes Dataset (FAD), having roughly 200,000 attribute labels for the above traits, for over 10,000 facial images. Due to the recent surge of research on Deep Convolutional Neural Networks (CNNs), we begin by using a CNN architecture for estimating facial attributes and show that they indeed provide an impressive baseline performance. To further improve performance, we propose a novel approach that incorporates facial landmark information for input images as an additional channel, helping the CNN learn better attribute-specific features so that the landmarks across various training images hold correspondence. We empirically analyse the performance of our method, showing consistent improvement over the baseline across traits.
【Keywords】:
【Paper Link】 【Pages】:1683-1689
【Authors】: Huayu Li ; Richang Hong ; Defu Lian ; Zhiang Wu ; Meng Wang ; Yong Ge
【Abstract】: Implicit feedback based recommendation has recently been an important task with the accumulated user-item interaction data. However, it is very challenging to produce recommendations from implicit feedback due to the sparseness of data and the lack of negative feedback/rating. Although various factor models have been proposed to tackle this problem, they either focus on rating prediction that may lead to inaccurate top-k recommendations or are dependent on the sampling of negative feedback that often results in bias. To this end, we propose a Relaxed Ranking-based Factor Model, RRFM, to relax pairwise ranking into a SVM-like task, where positive and negative feedbacks are separated by the soft boundaries, and their non-separate property is employed to capture the characteristic of unobserved data. A smooth and scalable algorithm is developed to solve group- and instance- level's optimization and parameter estimation. Extensive experiments based on real-world datasets demonstrate the effectiveness and advantage of our approach.
【Keywords】:
【Paper Link】 【Pages】:1690-1696
【Authors】: Jia Li ; Kaiser Asif ; Hong Wang ; Brian D. Ziebart ; Tanya Y. Berger-Wolf
【Abstract】: Providing sequence tagging that minimize Hamming loss is a challenging, but important, task. Directly minimizing this loss over a training sample is generally an NP-hard problem. Instead, existing sequence tagging methods minimize a convex upper bound that upper bounds the Hamming loss. Unfortunately, this often either leads to inconsistent predictors (e.g., max-margin methods) or predictions that are mismatched on the Hamming loss (e.g., conditional random fields). We present adversarial sequence tagging, a consistent structured prediction framework for minimizing Hamming loss by pessimistically viewing uncertainty. Our approach pessimistically approximates the training data, yielding an adversarial game between the sequence tag predictor and the sequence labeler. We demonstrate the benefits of the approach on activity recognition and information extraction/segmentation tasks.
【Keywords】:
【Paper Link】 【Pages】:1697-1703
【Authors】: Jingjing Li ; Jidong Zhao ; Ke Lu
【Abstract】: The essence of domain adaptation is to explore common latent factors shared by the involved domains. These factors can be specific features or geometric structures. Most of previous methods exploit either the shared features or the shared geometric structures separately. However, the two strategies are complementary with each other and jointly exploring them is more optimal. This paper proposes a novel approach, named joint Feature Selection and Structure Preservation (FSSP), for unsupervised domain adaptation. FSSP smoothly integrates structure preservation and feature selection into a unified optimization problem. Intensive experiments on text categorization, image classification and video event recognition demonstrate that our method performs better, even with up to 30% improvement in average, compared with the state-of-the-art methods.
【Keywords】:
【Paper Link】 【Pages】:1704-1710
【Authors】: Miaomiao Li ; Xinwang Liu ; Lei Wang ; Yong Dou ; Jianping Yin ; En Zhu
【Abstract】: Kernel alignment has recently been employed for multiple kernel clustering (MKC). However, we find that most of existing works implement this alignment in a global manner, which: i) indiscriminately forces all sample pairs to be equally aligned with the same ideal similarity; and ii) is inconsistent with a well-established concept that the similarity evaluated for two farther samples in a high dimensional space is less reliable. To address these issues, this paper proposes a novel MKC algorithm with a "local" kernel alignment, which only requires that the similarity of a sample to its k-nearest neighbours be aligned with the ideal similarity matrix. Such an alignment helps the clustering algorithm to focus on closer sample pairs that shall stay together and avoids involving unreliable similarity evaluation for farther sample pairs. We derive a new optimization problem to implement this idea, and design a two-step algorithm to efficiently solve it. As experimentally demonstrated on six challenging multiple kernel learning benchmark data sets, our algorithm significantly outperforms the state-of-the-art comparable methods in the recent literature, verifying the effectiveness and superiority of maximizing local kernel alignment.
【Keywords】:
【Paper Link】 【Pages】:1711-1717
【Authors】: Wu-Jun Li ; Sheng Wang ; Wang-Cheng Kang
【Abstract】: Recent years have witnessed wide application of hashing for large-scale image retrieval. However, most existing hashing methods are based on hand-crafted features which might not be optimally compatible with the hashing procedure. Recently, deep hashing methods have been proposed to perform simultaneous feature learning and hash-code learning with deep neural networks, which have shown better performance than traditional hashing methods with hand-crafted features. Most of these deep hashing methods are supervised whose supervised information is given with triplet labels. For another common application scenario with pairwise labels, there have not existed methods for simultaneous feature learning and hash-code learning. In this paper, we propose a novel deep hashing method, called deep pairwise-supervised hashing (DPSH), to perform simultaneous feature learning and hash-code learning for applications with pairwise labels. Experiments on real datasets show that our DPSH method can outperform other methods to achieve the state-of-the-art performance in image retrieval applications.
【Keywords】:
【Paper Link】 【Pages】:1718-1724
【Authors】: Yingming Li ; Ming Yang ; Zenglin Xu ; Zhongfei (Mark) Zhang
【Abstract】: Multi-view tagging has become increasingly popular in the applications where data representations by multiple views exist. A robust multi-view tagging method must have the capability to meet the two challenging requirements: limited labeled training samples and noisy labeled training samples. In this paper, we investigate this challenging problem of learning with limited and noisy tagging and propose a discriminative model, called MSMC, that exploits both labeled and unlabeled data through a semi-parametric regularization and takes advantage of the multi-label space consistency into the optimization. While MSMC is a general method for learning with multi-view, limited, and noisy tagging, in the evaluations we focus on the specific application of noisy image tagging with limited labeled training samples on a benchmark dataset. Extensive evaluations in comparison with state-of-the-art literature demonstrate that MSMC outstands with a superior performance.
【Keywords】:
【Paper Link】 【Pages】:1725-1731
【Authors】: Yu-Feng Li ; Shao-Bo Wang ; Zhi-Hua Zhou
【Abstract】: Graph as a common structure of machine learning, has played an important role in many learning tasks such as graph-based semi-supervised learning (GSSL). The quality of graph, however, seriously affects the performance of GSSL; moreover, an inappropriate graph may even cause deteriorated performance, that is, GSSL using unlabeled data may be outperformed by direct supervised learning with only labeled data. To this end, it is desired to judge the quality of graph and develop performance-safe GSSL methods. In this paper we propose a large margin separation method 'Lead' for safe GSSL. Our basic idea is that, if a certain graph owns a high quality, its predictive results on unlabeled data may have a large margin separation. We should exploit the large margin graphs while keeping the small margin graphs (which might be risky) to be rarely exploited. Based on this recognition, we formulate safe GSSL as Semi-Supervised SVM (S3VM) optimization and present an efficient algorithm. Extensive experimental results demonstrate that our proposed method can effectively improve the safeness of GSSL, in addition achieve highly competitive accuracy with many state-of-the-art GSSL methods.
【Keywords】:
【Paper Link】 【Pages】:1732-1738
【Authors】: Defu Lian ; Yong Ge ; Nicholas Jing Yuan ; Xing Xie ; Hui Xiong
【Abstract】: The popularity of social media creates a large amount of user-generated content, playing an important role in addressing cold-start problems in recommendation. Although much effort has been devoted to incorporating this information into recommendation, past work mainly targets explicit feedback. There is still no general framework tailored to implicit feedback, such as views, listens, or visits. To this end, we propose a sparse Bayesian content-aware collaborative filtering framework especially for implicit feedback, and develop a scalable optimization algorithm to jointly learn latent factors and hyperparameters. Due to the adaptive update of hyperparameters, automatic feature selection is naturally embedded in this framework. Convincing experimental results on three different implicit feedback datasets indicate the superiority of the proposed algorithm to state-of-the-art content-aware recommendation methods.
【Keywords】:
【Paper Link】 【Pages】:1739-1745
【Authors】: Jian Liang ; Ran He ; Zhenan Sun ; Tieniu Tan
【Abstract】: Cross-modal learning tries to find various types of heterogeneous data (e.g., image) from a given query (e.g., text). Most cross-modal algorithms heavily rely on semantic labels and benefit from a semantic-preserving aggregation of pairs of heterogeneous data. However, the semantic labels are not readily obtained in many real-world applications. This paper studies the aggregation of these pairs unsupervisedly. Apart from lower pairwise correspondences that force the data from one pair to be close to each other, we propose a novel concept, referred as groupwise correspondences, supposing that each paired heterogeneous data are from an identical latent group. We incorporate this groupwise correspondences into canonical correlation analysis (CCA) model, and seek a latent common subspace where data are naturally clustered into several latent groups. To simplify this nonconvex and nonsmooth problem, we introduce a non-negative orthogonal variable to represent the soft group membership, then two coupled computationally efficient subproblems (a generalized ratio-trace problem and a non-negative problem) are alternatively minimized to guarantee the proposed algorithm converges locally. Experimental results on two benchmark datasets demonstrate that the proposed unsupervised algorithm even achieves comparable performance to some state-of-the-art supervised cross-modal algorithms.Cross-modal learning tries to find various types of heterogeneous data (e.g., image) from a given query (e.g., text). Most cross-modal algorithms heavily rely on semantic labels and benefit from a semantic-preserving aggregation of pairs of heterogeneous data. However, the semantic labels are not readily obtained in many real-world applications. This paper studies the aggregation of these pairs unsupervisedly. Apart from lower pairwise correspondences that force the data from one pair to be close to each other, we propose a novel concept, referred as groupwise correspondences, supposing that each paired heterogeneous data are from an identical latent group. We incorporate this groupwise correspondences into canonical correlation analysis (CCA) model, and seek a latent common subspace where data are naturally clustered into several latent groups. To simplify this nonconvex and nonsmooth problem, we introduce a non-negative orthogonal variable to represent the soft group membership, then two coupled computationally efficient subproblems (a generalized ratio-trace problem and a non-negative problem) are alternatively minimized to guarantee the proposed algorithm converges locally. Experimental results on two benchmark datasets demonstrate that the proposed unsupervised algorithm even achieves comparable performance to some state-of-the-art supervised cross-modal algorithms.
【Keywords】:
【Paper Link】 【Pages】:1746-1752
【Authors】: Junwei Liang ; Lu Jiang ; Deyu Meng ; Alexander G. Hauptmann
【Abstract】: Learning detectors that can recognize concepts, such as people actions, objects, etc., in video content is an interesting but challenging problem. In this paper, we study the problem of automatically learning detectors from the big video data on the web without any additional manual annotations. The contextual information available on the web provides noisy labels to the video content. To leverage the noisy web labels, we propose a novel method called WEbly-Labeled Learning (WELL). It is established on two theories called curriculum learning and self-paced learning and exhibits useful properties that can be theoretically verified. We provide compelling insights on the latent non-convex robust loss that is being minimized on the noisy data. In addition, we propose two novel techniques that not only enable WELL to be applied to big data but also lead to more accurate results. The efficacy and the scalability of WELL have been extensively demonstrated on two public benchmarks, including the largest multimedia dataset and the largest manually-labeled video set. Experimental results show that WELL significantly outperforms the state-of-the-art methods. To the best of our knowledge, WELL achieves by far the best reported performance on these two webly-labeled big video datasets.
【Keywords】:
【Paper Link】 【Pages】:1753-1759
【Authors】: Shaohui Lin ; Rongrong Ji ; Xiaowei Guo ; Xuelong Li
【Abstract】: In recent years, convolutional neural networks (CNNs) have achieved remarkable success in various applications such as image classification, object detection, object parsing and face alignment. Such CNN models are extremely powerful to deal with massive amounts of training data by using millions and billions of parameters. However, these models are typically deficient due to the heavy cost in model storage, which prohibits their usage on resource-limited applications like mobile or embedded devices. In this paper, we target at compressing CNN models to an extreme without significantly losing their discriminability. Our main idea is to explicitly model the output reconstruction error between the original and compressed CNNs, which error is minimized to pursuit a satisfactory rate-distortion after compression. In particular, a global error reconstruction method termed GER is presented, which firstly leverages an SVD-based low-rank approximation to coarsely compress the parameters in the fully connected layers in a layer-wise manner. Subsequently, such layer-wise initial compressions are jointly optimized in a global perspective via back-propagation. The proposed GER method is evaluated on the ILSVRC2012 image classification benchmark, with implementations on two widely-adopted convolutional neural networks, i.e. the AlexNet and VGGNet-19. Comparing to several state-of-the-art and alternative methods of CNN compression, the proposed scheme has demonstrated the best rate-distortion performance on both networks.
【Keywords】:
【Paper Link】 【Pages】:1760-1766
【Authors】: Bo Liu ; Xiao-Tong Yuan ; Shaoting Zhang ; Qingshan Liu ; Dimitris N. Metaxas
【Abstract】: The k-support-norm regularized minimization has recently been applied with success to sparse prediction problems. The proximal gradient method is conventionally used to minimize this composite model. However it tends to suffer from expensive iteration cost as the proximity operator associated with k-support-norm needs exhaustive searching operations and thus could be time consuming in large scale settings. In this paper, we reformulate the k-support-norm regularized formulation into an identical constrained formulation and propose a fully corrective Frank-Wolfe algorithm to minimize the constrained model. Our method is inspired by an interesting observation that the convex hull structure of the k-support-norm ball allows the application of Frank-Wolfe-type algorithms with low iteration complexity. The convergence behavior of the proposed algorithm is analyzed. Extensive numerical results in learning tasks including logistic regression and matrix pursuit demonstrate the substantially improved computational efficiency of our algorithm over the state-of-the-art proximal gradient algorithms.
【Keywords】:
【Paper Link】 【Pages】:1767-1773
【Authors】: Hong Liu ; Rongrong Ji ; Yongjian Wu ; Gang Hua
【Abstract】: Matrix factorization has been recently utilized for the task of multi-modal hashing for cross-modality visual search, where basis functions are learned to map data from different modalities to the same Hamming embedding. In this paper, we propose a novel cross-modality hashing algorithm termed Supervised Matrix Factorization Hashing (SMFH) which tackles the multi-modal hashing problem with a collective non-negative matrix factorization across the different modalities. In particular, SMFH employs a well-designed binary code learning algorithm to preserve the similarities among multi-modal original features through a graph regularization. At the same time, semantic labels, when available, are incorporated into the learning procedure.We conjecture that all these would facilitate to preserve the most relevant information during the binary quantization process, and hence improve the retrieval accuracy. We demonstrate the superior performance of SMFH on three cross-modality visual search benchmarks, i.e., the PASCAL-Sentence, Wiki, and NUS-WIDE, with quantitative comparison to various state-of-the-art methods.
【Keywords】:
【Paper Link】 【Pages】:1774-1780
【Authors】: Li Liu ; William K. Cheung ; Xin Li ; Lejian Liao
【Abstract】: In this paper, we adopt the representation learning approach to align users across multiple social networks where the social structures of the users are exploited. In particular, we propose to learn a network embedding with the follower-ship/followee-ship of each user explicitly modeled as input/output context vector representations so as to preserve the proximity of users with "similar" followers/followees in the embedded space. For the alignment, we add both known and potential anchor users across the networks to facilitate the transfer of context information across networks. We solve both the network embedding problem and the user alignment problem simultaneously under a unified optimization framework. The stochastic gradient descent and negative sampling algorithms are used to address scalability issues. Extensive experiments on real social network datasets demonstrate the effectiveness and efficiency of the proposed approach compared with several state-of-the-art methods.
【Keywords】:
【Paper Link】 【Pages】:1781-1787
【Authors】: Li-Ping Liu ; Thomas G. Dietterich ; Nan Li ; Zhi-Hua Zhou
【Abstract】: Consider a binary classification problem in which the learner is given a labeled training set, an unlabeled test set, and is restricted to choosing exactly k test points to output as positive predictions. Problems of this kind — transductive precision@k — arise in many applications. Previous methods solve these problems in two separate steps, learning the model and selecting k test instances by thresholding their scores. In this way, model training is not aware of the constraint of choosing k test instances as positive in the test phase. This paper shows the importance of incorporating the knowledge of k into the learning process and introduces a new approach, Transductive Top K (TTK), that seeks to minimize the hinge loss over all training instances under the constraint that exactly k test instances are predicted as positive. The paper presents two optimization methods for this challenging problem. Experiments and analysis confirm the benefit of incoporating k in the learning process. In our experimental evaluations, the performance of TTK matches or exceeds existing state-of-the-art methods on 7 benchmark datasets for binary classification and 3 reserve design problem instances.
【Keywords】:
【Paper Link】 【Pages】:1788-1794
【Authors】: Qi Liu ; Hongtao Lu
【Abstract】: Among learning-based hashing methods, supervised hashing tries to find hash codes which preserve semantic similarities of original data. Recent years have witnessed much efforts devoted to design objective functions and optimization methods for supervised hashing learning, in order to improve search accuracy and reduce training cost. In this paper, we propose a very straightforward supervised hashing algorithm and demonstrate its superiority over several state-of-the-art methods. The key idea of our approach is to treat label vectors as binary codes and to learn target codes which have similar structure to label vectors. To circumvent direct optimization on large Gram matrices, we identify an inner-product-preserving transformation and use it to bring close label vectors and hash codes without changing the structure. The optimization process is very efficient and scales well. In our experiment, training 16-bit and 96-bit code on NUS-WIDE cost respectively only 3 and 6 minutes.
【Keywords】:
【Paper Link】 【Pages】:1795-1801
【Authors】: Mario Lucic ; Olivier Bachem ; Andreas Krause
【Abstract】: Outliers are ubiquitous in modern data sets. Distance-based techniques are a popular non-parametric approach to outlier detection as they require no prior assumptions on the data generating distribution and are simple to implement. Scaling these techniques to massive data sets without sacrificing accuracy is a challenging task. We propose a novel algorithm based on the intuition that outliers have a significant influence on the quality of divergence-based clustering solutions. We propose sensitivity - the worst-case impact of a data point on the clustering objective - as a measure of outlierness. We then prove that influence - a (non-trivial) upper-bound on the sensitivity can be computed by a simple linear time algorithm. To scale beyond a single machine, we propose a communication efficient distributed algorithm. In an extensive experimental evaluation, we demonstrate the effectiveness and establish the statistical significance of the proposed approach. In particular, it outperforms the most popular distance-based approaches while being several orders of magnitude faster.
【Keywords】:
【Paper Link】 【Pages】:1802-1808
【Authors】: Minnan Luo ; Feiping Nie ; Xiaojun Chang ; Yi Yang ; Alexander G. Hauptmann ; Qinghua Zheng
【Abstract】: Robust principal component analysis (PCA) is one of the most important dimension reduction techniques to handle high-dimensional data with outliers. However, the existing robust PCA presupposes that the mean of the data is zero and incorrectly utilizes the Euclidean distance based optimal mean for robust PCA with ℓ1-norm. Some studies consider this issue and integrate the estimation of the optimal mean into the dimension reduction objective, which leads to expensive computation.In this paper, we equivalently reformulate the maximization of variances for robust PCA, such that the optimal projection directions are learned by maximizing the sum of the projected difference between each pair of instances, rather than the difference between each instance and the mean of the data.Based on this reformulation, we propose a novel robust PCA to automatically avoid the calculation of the optimal mean based on ℓ1-norm distance. This strategy also makes the assumption of centered data unnecessary. Additionally, we intuitively extend the proposed robust PCA to its 2D version for image recognition. Efficient non-greedy algorithms are exploited to solve the proposed robust PCA and 2D robust PCA with fast convergence and low computational complexity.Some experimental results on benchmark data sets demonstrate the effectiveness and superiority of the proposed approaches on image reconstruction and recognition.
【Keywords】:
【Paper Link】 【Pages】:1809-1815
【Authors】: Yong Luo ; Yonggang Wen ; Dacheng Tao
【Abstract】: Distance metric learning (DML) is critical for a wide variety of machine learning algorithms and pattern recognition applications. Transfer metric learning (TML) leverages the side information (e.g., similar/dissimilar constraints over pairs of samples) from related domains to help the target metric learning (with limited information). Current TML tools usually assume that different domains exploit the same feature representation, and thus are not applicable to tasks where data are drawn from heterogeneous domains. Heterogeneous transfer learning approaches handle heterogeneous domains by usually learning feature transformations across different domains. The learned transformation can be used to derive a metric, but these approaches are mostly limited by their capability of only handling two domains. This motivates the proposed heterogeneous multi-task metric learning (HMTML) framework for handling multiple domains by combining side information and unlabeled data. Specifically, HMTML learns the metrics for all different domains simultaneously by maximizing their high-order correlation (parameterized by feature covariance of unlabeled data) in a common subspace, which is induced by the transformations derived from the metrics. Extensive experiments on both multi-language text categorization and multi-view social image annotation demonstrate the effectiveness of the proposed method.
【Keywords】:
【Paper Link】 【Pages】:1816-1822
【Authors】: Weizhi Ma ; Min Zhang ; Yiqun Liu ; Shaoping Ma
【Abstract】: Large-scale customer service call records include lots of valuable information for business intelligence. However, the analysis of those records has not utilized in the big data era before. There are two fundamental problems before mining and analyses: 1) The telephone conversation is mixed with words of agents and users which have to be recognized before analysis; 2) The speakers in conversation are not in a pre-defined set. These problems are new challenges which have not been well studied in the previous work. In this paper, we propose a four-phase framework for role labeling in real customer service telephone conversation, with the benefit of integrating multi-modality features, i.e., both low-level acoustic features and semantic-level textual features. Firstly, we conduct ΔBayesian Information Criterion (ΔBIC) based speaker diarization to get two segments clusters from an audio stream. Secondly, the segments are transferred into text in an Automatic Speech Recognition (ASR) phase with a deep learning model DNN-HMM. Thirdly, by integrating acoustic and textual features, dialog level role labeling is proposed to map the two clusters into the agent and the user. Finally, sentence level role correction is designed in order to label results correctly in a fine-grained notion, which reduces the errors made in previous phases. The proposed framework is tested on two real datasets: mobile and bank customer service calls datasets. The precision of dialog level labeling is over 99.0%. On the sentence level, the accuracy of labeling reaches 90.4%, greatly outperforming traditional acoustic features based method which achieves only 78.5% in accuracy.
【Keywords】:
【Paper Link】 【Pages】:1823-1829
【Authors】: Tong Man ; Huawei Shen ; Shenghua Liu ; Xiaolong Jin ; Xueqi Cheng
【Abstract】: Predicting anchor links across social networks has important implications to an array of applications, including cross-network information diffusion and cross-domain recommendation. One challenging problem is: whether and to what extent we can address the anchor link prediction problem, if only structural information of networks is available. Most existing methods, unsupervised or supervised, directly work on networks themselves rather than on their intrinsic structural regularities, and thus their effectiveness is sensitive to the high dimension and sparsity of networks. To offer a robust method, we propose a novel supervised model, called PALE, which employs network embedding with awareness of observed anchor links as supervised information to capture the major and specific structural regularities and further learns a stable cross-network mapping for predicting anchor links. Through extensive experiments on two realistic datasets, we demonstrate that PALE significantly outperforms the state-of-the-art methods.
【Keywords】:
【Paper Link】 【Pages】:1830-1838
【Authors】: Travis Mandel ; Yun-En Liu ; Emma Brunskill ; Zoran Popovic
【Abstract】: A fundamental artificial intelligence challenge is how to design agents that intelligently trade off exploration and exploitation while quickly learning about an unknown environment. However, in order to learn quickly, we must somehow generalize experience across states. One promising approach is to use Bayesian methods to simultaneously cluster dynamics and control exploration; unfortunately, these methods tend to require computationally intensive MCMC approximation techniques which lack guarantees. We propose Thompson Clustering for Reinforcement Learning (TCRL), a family of Bayesian clustering algorithms for reinforcement learning that leverage structure in the state space to remain computationally efficient while controlling both exploration and generalization. TCRL-Theoretic achieves near-optimal Bayesian regret bounds while consistently improving over a standard Bayesian exploration approach. TCRL-Relaxed is guaranteed to converge to acting optimally, and empirically outperforms state-of-the-art Bayesian clustering algorithms across a variety of simulated domains, even in cases where no states are similar.
【Keywords】:
【Paper Link】 【Pages】:1839-1845
【Authors】: Liang Mao ; Shiliang Sun
【Abstract】: Multi-view learning receives increasing interest in recent years to analyze complex data. Lately, multi-view maximum entropy discrimination (MVMED) and alternative MVMED (AMVMED) were proposed as extensions of maximum entropy discrimination (MED) to the multi-view learning setting, which use the hard margin consistency principle that enforces two view margins to be the same. In this paper, we propose soft margin consistency based multi-view MED (SMVMED) achieving margin consistency in a less strict way, which minimizes the relative entropy between the posteriors of two view margins. With a trade-off parameter balancing large margin and margin consistency, SMVMED is more flexible. We also propose a sequential minimal optimization (SMO) algorithm to efficiently train SMVMED and make it scalable to large datasets. We evaluate the performance of SMVMED on multiple real-world datasets and get encouraging results.
【Keywords】:
【Paper Link】 【Pages】:1846-1852
【Authors】: Mazen Melibari ; Pascal Poupart ; Prashant Doshi
【Abstract】: Investigations into probabilistic graphical models for decision making have predominantly centered on influence diagrams (IDs) and decision circuits (DCs) for representation and computation of decision rules that maximize expected utility. Since IDs are typically handcrafted and DCs are compiled from IDs, in this paper we propose an approach to learn the structure and parameters of decision-making problems directly from data. We present a new representation called sum-product-max network (SPMN) that generalizes a sum-product network (SPN) to the class of decision-making problems and whose solution, analogous to DCs, scales linearly in the size of the network. We show that SPMNs may be reduced to DCs linearly and present a first method for learning SPMNs from data. This approach is significant because it facilitates a novel paradigm of tractable decision making driven by data.
【Keywords】:
【Paper Link】 【Pages】:1853-1859
【Authors】: Qi Meng ; Wei Chen ; Jingcheng Yu ; Taifeng Wang ; Zhiming Ma ; Tie-Yan Liu
【Abstract】: Stochastic gradient descent (SGD) is a widely used optimization algorithm in machine learning. In order to accelerate the convergence of SGD, a few advanced techniques have been developed in recent years, including variance reduction, stochastic coordinate sampling, and Nesterov's acceleration method. Furthermore, in order to improve the training speed and/or leverage larger-scale training data, asynchronous parallelization of SGD has also been studied. Then, a natural question is whether these techniques can be seamlessly integrated with each other, and whether the integration has desirable theoretical guarantee on its convergence. In this paper, we provide our formal answer to this question. In particular, we consider the asynchronous parallelization of SGD, accelerated by leveraging variance reduction, coordinate sampling, and Nesterov's method. We call the new algorithm asynchronous accelerated SGD (AASGD).Theoretically, we proved a convergence rate of AASGD, which indicates that (i) the three acceleration methods are complementary to each other and can make their own contributions to the improvement of convergence rate; (ii) asynchronous parallelization does not hurt the convergence rate,and can achieve considerable speedup under appropriate parameter setting. Empirically, we tested AASGD on a few benchmark datasets. The experimental results verified our theoretical findings and indicated that AASGD could be a highly effective and efficient algorithm for practical use.
【Keywords】:
【Paper Link】 【Pages】:1860-1866
【Authors】: Yadong Mu ; Wei Liu ; Cheng Deng ; Zongting Lv ; Xinbo Gao
【Abstract】: Learning compact hash codes has been a vibrant research topic for large-scale similarity search owing to the low storage cost and expedited search operation. A recent research thrust aims to learn compact codes jointly from multiple sources, referred to as cross-view (or cross-modal) hashing in the literature. The main theme of this paper is to develop a novel formulation and optimization scheme for cross-view hashing. As a key differentiator, our proposed method directly conducts optimization on discrete binary hash codes, rather than relaxed continuous variables as in existing cross-view hashing methods. This way relaxation-induced search accuracy loss can be avoided. We attack the cross-view hashing problem by simultaneously capturing semantic neighboring relations and maximizing the generative probability of the learned hash codes in each view. Specifically, to enable effective optimization on discrete hash codes, the optimization proceeds in a block coordinate descent fashion. Each iteration sequentially updates a single bit with others clamped. We transform the resultant sub-problem into an equivalent, more tractable quadratic form and devise an active set based solver on the discrete codes. Rigorous theoretical analysis is provided for the convergence and local optimality condition. Comprehensive evaluations are conducted on three image benchmarks. The clearly superior experimental results faithfully prove the merits of the proposed method.
【Keywords】:
【Paper Link】 【Pages】:1867-1973
【Authors】: Syed Abbas Z. Naqvi ; Shandian Zhe ; Yuan Qi ; Yifan Yang ; Jieping Ye
【Abstract】: We consider the application of Bayesian spike-and-slab models in high-dimensional feature selection problems. To do so, we propose a simple yet effective fast approximate Bayesian inference algorithm based on Laplace's method. We exploit two efficient optimization methods, GIST and L-BFGS, to obtain the mode of the posterior distribution. Then we propose an ensemble Nystrom based approach to calculate the diagonal of the inverse Hessian over the mode to obtain the approximate posterior marginals in O(knp) time, k ≪ p. Furthermore, we provide the theoretical analysis about the estimation consistency and approximation error bounds. With the posterior marginals of the model weights, we use quadrature integration to estimate the marginal posteriors of selection probabilities and indicator variables for all features, which quantify the selection uncertainty. Our method not only maintains the benefits of the Bayesian treatment (e.g., uncertainty quantification) but also possesses the computational efficiency, and oracle properties of the frequentist methods. Simulation shows that our method estimates better or comparable selection probabilities and indicator variables than alternative approximate inference methods such as VB and EP, but with less running time. Extensive experiments on large real datasets demonstrate that our method often improves prediction accuracy over Bayesian automatic relevance determination, EP, and frequentist L1 type methods.
【Keywords】:
【Paper Link】 【Pages】:1874-1880
【Authors】: Feiping Nie ; Heng Huang
【Abstract】: We propose a new subspace clustering model to segment data which is drawn from multiple linear or affine subspaces. Unlike the well-known sparse subspace clustering (SSC) and low-rank representation (LRR) which transfer the subspace clustering problem into two steps' algorithm including building the affinity matrix and spectral clustering, our proposed model directly learns the different subspaces' indicator so that low-rank based different groups are obtained clearly. To better approximate the low-rank constraint, we suggest to use Schatten p-norm to relax the rank constraint instead of using trace norm. We tactically avoid the integer programming problem imposed by group indicator constraint to let our algorithm more efficient and scalable. Furthermore, we extend our discussion to the general case in which subspaces don't pass the original point. The new algorithm's convergence is given, and both synthetic and real world datasets demonstrate our proposed model's effectiveness.
【Keywords】:
【Paper Link】 【Pages】:1881-1887
【Authors】: Feiping Nie ; Jing Li ; Xuelong Li
【Abstract】: Graph-based approaches have been successful in unsupervised and semi-supervised learning. In this paper, we focus on the real-world applications where the same instance can be represented by multiple heterogeneous features. The key point of utilizing the graph-based knowledge to deal with this kind of data is to reasonably integrate the different representations and obtain the most consistent manifold with the real data distributions. In this paper, we propose a novel framework via the reformulation of the standard spectral learning model, which can be used for multiview clustering and semi-supervised tasks. Unlike other methods in the literature, the proposed method can learn an optimal weight for each graph automatically without introducing an additive parameter as previous methods do. Furthermore, our objective under semi-supervised learning is convex and the global optimal result will be obtained. Extensive empirical results on different real-world data sets demonstrate that the proposed method achieves comparable performance with the state-of-the-art approaches and can be used more practically.
【Keywords】:
【Paper Link】 【Pages】:1888-1894
【Authors】: Shohei Ohsawa ; Yachiko Obara ; Takayuki Osogami
【Abstract】: Recommender systems rely on techniques of predicting the ratings that users would give to yet unconsumed items. Probabilistic matrix factorization (PMF) is a standard technique for such prediction and makes a prediction on the basis of an underlying probabilistic generative model of the behavior of users. We investigate a new model of users' consumption and rating, where a user tends to consume an item that emphasizes those features that the user seeks to enjoy, and the ratings of the users are more strongly affected by those features than others. We incorporate this new user model into PMF and show that the resulting method, Gated PMF (GPMF), improves the predictive accuracy by several percent on standard datasets. GPMF is widely applicable, as it is trained only with the ratings given by users and does not rely on any auxiliary data.
【Keywords】:
【Paper Link】 【Pages】:1895-1901
【Authors】: Shirui Pan ; Jia Wu ; Xingquan Zhu ; Chengqi Zhang ; Yang Wang
【Abstract】: Information network mining often requires examination of linkage relationships between nodes for analysis. Recently, network representation has emerged to represent each node in a vector format, embedding network structure, so off-the-shelf machine learning methods can be directly applied for analysis. To date, existing methods only focus on one aspect of node information and cannot leverage node labels. In this paper, we propose TriDNR, a tri-party deep network representation model, using information from three parties: node structure, node content, and node labels (if available) to jointly learn optimal node representation. TriDNR is based on our new coupled deep natural language module, whose learning is enforced at three levels: (1) at the network structure level, TriDNR exploits inter-node relationship by maximizing the probability of observing surrounding nodes given a node in random walks; (2) at the node content level, TriDNR captures node-word correlation by maximizing the co-occurrence of word sequence given a node; and (3) at the node label level, TriDNR models label-word correspondence by maximizing the probability of word sequence given a class label. The tri-party information is jointly fed into the neural network model to mutually enhance each other to learn optimal representation, and results in up to 79% classification accuracy gain, compared to state-of-the-art methods.
【Keywords】:
【Paper Link】 【Pages】:1902-1908
【Authors】: Guansong Pang ; Longbing Cao ; Ling Chen
【Abstract】: This paper introduces a novel unsupervised outlier detection method, namely Coupled Biased Random Walks (CBRW), for identifying outliers in categorical data with diversified frequency distributions and many noisy features. Existing pattern-based outlier detection methods are ineffective in handling such complex scenarios, as they misfit such data. CBRW estimates outlier scores of feature values by modelling feature value level couplings, which carry intrinsic data characteristics, via biased random walks to handle this complex data. The outlier scores of feature values can either measure the outlierness of an object or facilitate the existing methods as a feature weighting and selection indicator. Substantial experiments show that CBRW can not only detect outliers in complex data significantly better than the state-of-the-art methods, but also greatly improve the performance of existing methods on data sets with many noisy features.
【Keywords】:
【Paper Link】 【Pages】:1909-1917
【Authors】: Giorgio Patrini ; Richard Nock ; Stephen Hardy ; Tiberio Caetano
【Abstract】: Consider the following scenario: two datasets/peers contain the same real-world entities described using partially shared features, e.g. banking and insurance company records of the same customer base. Our goal is to learn a classifier in the cross product space of the two domains, in the hard case in which no shared ID is available — e.g. due to anonymization. Traditionally, the problem is approached by first addressing entity matching and subsequently learning the classifier in a standard manner. We present an end-to-end solution which bypasses matching entities, based on the recently introduced concept of Rademacher observations (rados). Informally, we replace the minimisation of a loss over examples, which requires entity resolution, by the equivalent minimisation of a (different) loss over rados. We show that (i) a potentially exponential-size subset of these rados does not require to perform entity matching, and (ii) the algorithm that provably minimizes the loss over rados has time and space complexities smallerthan the algorithm minimizing the equivalent example loss.Last, we relax a key assumption, that the data is vertically partitioned among peers — in this case, we would not even know the existence of a solution to entity resolution. In this more general setting, experiments validate the possibility of beating even the optimal peer in hindsight.
【Keywords】:
【Paper Link】 【Pages】:1918-1924
【Authors】: Hanyang Peng ; Yong Fan
【Abstract】: A novel sparsity optimization method is proposed to select features for multi-class classification problems by directly optimizing a ℓ2,p-norm (0 < p ≤ 1) based sparsity function subject to data-fitting inequality constraints to obtain large between-class margins. The direct sparse optimization method circumvents the empirical tuning of regularization parameters in existing feature selection methods that adopt the sparsity model as a regularization term. To solve the direct sparsity optimization problem that is non-smooth and non-convex when 0 < p < 1, we propose an efficient iterative algorithm with proved convergence by converting it to a convex and smooth optimization problem at every iteration step. The proposed algorithm has been evaluated based on publicly available datasets. The experiments have demonstrated that our algorithm could achieve feature selection performance competitive to state-of-the-art algorithms.
【Keywords】:
【Paper Link】 【Pages】:1925-1931
【Authors】: Xi Peng ; Shijie Xiao ; Jiashi Feng ; Wei-Yun Yau ; Zhang Yi
【Abstract】: Subspace clustering aims to cluster unlabeled samples into multiple groups by implicitly seeking a subspace to fit each group. Most of existing methods are based on a shallow linear model, which may fail in handling data with nonlinear structure. In this paper, we propose a novel subspace clustering method — deeP subspAce clusteRing with sparsiTY prior (PARTY) — based on a new deep learning architecture. PARTY explicitly learns to progressively transform input data into nonlinear latent space and to be adaptive to the local and global subspace structure simultaneously. In particular, considering local structure, PARTY learns representation for the input data with minimal reconstruction error. Moreover, PARTY incorporates a prior sparsity information into the hidden representation learning to preserve the sparse reconstruction relation over the whole data set. To the best of our knowledge, PARTY is the first deep learning based subspace clustering method. Extensive experiments verify the effectiveness of our method.
【Keywords】:
【Paper Link】 【Pages】:1932-1938
【Authors】: Te Pi ; Xi Li ; Zhongfei Zhang ; Deyu Meng ; Fei Wu ; Jun Xiao ; Yueting Zhuang
【Abstract】: Effectiveness and robustness are two essential aspects of supervised learning studies. For effective learning, ensemble methods are developed to build a strong effective model from ensemble of weak models. For robust learning, self-paced learning (SPL) is proposed to learn in a self-controlled pace from easy samples to complex ones. Motivated by simultaneously enhancing the learning effectiveness and robustness, we propose a unified framework, Self-Paced Boost Learning (SPBL). With an adaptive from-easy-to-hard pace in boosting process, SPBL asymptotically guides the model to focus more on the insufficiently learned samples with higher reliability. Via a max-margin boosting optimization with self-paced sample selection, SPBL is capable of capturing the intrinsic inter-class discriminative patterns while ensuring the reliability of the samples involved in learning. We formulate SPBL as a fully-corrective optimization for classification. The experiments on several real-world datasets show the superiority of SPBL in terms of both effectiveness and robustness.
【Keywords】:
【Paper Link】 【Pages】:1939-1945
【Authors】: Chao Qian ; Jing-Cheng Shi ; Yang Yu ; Ke Tang ; Zhi-Hua Zhou
【Abstract】: Subset selection that selects a few variables from a large set is a fundamental problem in many areas. The recently emerged Pareto Optimization for Subset Selection (POSS) method is a powerful approximation solver for this problem. However, POSS is not readily parallelizable, restricting its large-scale applications on modern computing architectures. In this paper, we propose PPOSS, a parallel version of POSS. Our theoretical analysis shows that PPOSS has good properties for parallelization while preserving the approximation quality: when the number of processors is limited (less than the total number of variables), the running time of PPOSS can be reduced almost linearly with respect to the number of processors; with increasing number of processors, the running time can be further reduced, eventually to a constant. Empirical studies verify the effectiveness of PPOSS, and moreover suggest that the asynchronous implementation is more efficient with little quality loss.
【Keywords】:
【Paper Link】 【Pages】:1946-1952
【Authors】: Hong Qian ; Yi-Qi Hu ; Yang Yu
【Abstract】: Derivative-free optimization methods are suitable for sophisticated optimization problems, while are hard to scale to high dimensionality (e.g., larger than 1,000). Previously, the random embedding technique has been shown successful for solving high-dimensional problems with low effective dimensions. However, it is unrealistic to assume a low effective dimension in many applications. This paper turns to study high-dimensional problems with low optimal epsilon-effective dimensions, which allow all dimensions to be effective but many of them only have a small bounded effect. We characterize the properties of random embedding for this kind of problems, and propose the sequential random embeddings (SRE) to reduce the embedding gap while running optimization algorithms in the low-dimensional spaces. We apply SRE to several state-of-the-art derivative-free optimization methods, and conduct experiments on synthetic functions as well as non-convex classification tasks with up to 100,000 variables. Experiment results verify the effectiveness of SRE.
【Keywords】:
【Paper Link】 【Pages】:1953-1959
【Authors】: Peng Qian ; Xipeng Qiu ; Xuanjing Huang
【Abstract】: Recently, the long short-term memory neural network (LSTM) has attracted wide interest due to its success in many tasks. LSTM architecture consists of a memory cell and three gates, which looks similar to the neuronal networks in the brain. However, there still lacks the evidence of the cognitive plausibility of LSTM architecture as well as its working mechanism.In this paper, we study the cognitive plausibility of LSTM by aligning its internal architecture with the brain activity observed via fMRI when the subjects read a story. Experiment results show that the artificial memory vector in LSTM can accurately predict the observed sequential brain activities, indicating the correlation between LSTM architecture and the cognitive process of story reading.
【Keywords】:
【Paper Link】 【Pages】:1960-1966
【Authors】: Wei Qian ; Bin Hong ; Deng Cai ; Xiaofei He ; Xuelong Li
【Abstract】: Non-negative Matrix Factorization (NMF) has received considerable attentions in various areas for its psychological and physiological interpretation of naturally occurring data whose representation may be parts-based in the human brain. Despite its good practical performance, one shortcoming of original NMF is that it ignores intrinsic structure of data set. On one hand, samples might be on a manifold and thus one may hope that geometric information can be exploited to improve NMF's performance. On the other hand, features might correlate with each other, thus conventional L2 distance can not well measure the distance between samples. Although some works have been proposed to solve these problems, rare connects them together. In this paper, we propose a novel method that exploits knowledge in both data manifold and features correlation. We adopt an approximation of Earth Mover's Distance (EMD) as metric and add a graph regularized term based on EMD to NMF. Furthermore, we propose an efficient multiplicative iteration algorithm to solve it. Our empirical study shows the encouraging results of the proposed algorithm comparing with other NMF methods.
【Keywords】:
【Paper Link】 【Pages】:1967-1973
【Authors】: Vaibhav Rajan ; Sakyajit Bhattacharya
【Abstract】: Heterogeneous data with complex feature dependencies is common in real-world applications. Clustering algorithms for mixed - continuous and discrete valued - features often do not adequately model dependencies and are limited to modeling meta-Gaussian distributions. Copulas, that provide a modular parameterization of joint distributions, can model a variety of dependencies but their use with discrete data remains limited due to challenges in parameter inference. In this paper we use Gaussian mixture copulas, to model complex dependencies beyond those captured by meta-Gaussian distributions, for clustering. We design a new, efficient, semiparametric algorithm to approximately estimate the parameters of the copula that can fit continuous, ordinal and binary data. We analyze the conditions for obtaining consistent estimates and empirically demonstrate performance improvements over state-of-the-art methods of correlation clustering on synthetic and benchmark datasets.
【Keywords】:
【Paper Link】 【Pages】:1974-1982
【Authors】: Daniel P. Robinson ; Suchi Saria
【Abstract】: Predictive models are finding an increasing number of applications in many industries. As a result, a practical means for trading-off the cost of deploying a model versus its effectiveness is needed. Our work is motivated by risk prediction problems in healthcare. Cost-structures in domains such as healthcare are quite complex, posing a significant challenge to existing approaches. We propose a novel framework for designing cost-sensitive structured regularizers that is suitable for problems with complex cost dependencies. We draw upon a surprising connection to boolean circuits. In particular, we represent the problem costs as a multi-layer boolean circuit, and then use properties of boolean circuits to define an extended feature vector and a group regularizer that exactly captures the underlying cost structure. The resulting regularizer may then be combined with a fidelity function to perform model prediction, for example. For the challenging real-world application of risk prediction for sepsis in intensive care units, the use of our regularizer leads to models that are in harmony with the underlying cost structure and thus provide an excellent prediction accuracy versus cost tradeoff.
【Keywords】:
【Paper Link】 【Pages】:1983-1989
【Authors】: Weiwei Shen ; Jun Wang
【Abstract】: As a definitive investment guideline for institutions and individuals, Markowitz's modern portfolio theory is ubiquitous in financial industry. However, its noticeably poor out-of-sample performance due to the inaccurate estimation of parameters evokes unremitting efforts of investigating effective remedies. One common retrofit that blends portfolios from disparate investment perspectives has received growing attention. While even a naive portfolio blending strategy can be empirically successful, how to effectually and robustly blend portfolios to generate stable performance improvement remains less explored. In this paper, we present a novel online algorithm that leverages Thompson sampling into the sequential decision-making process for portfolio blending. By modeling blending coefficients as probabilities of choosing basis portfolios and utilizing Bayes decision rules to update the corresponding distribution functions, our algorithm sequentially determines the optimal coefficients to blend multiple portfolios that embody different criteria of investment and market views. Compared with competitive trading strategies across various benchmarks, our method shows superiority through standard evaluation metrics.
【Keywords】:
【Paper Link】 【Pages】:1990-1996
【Authors】: Zebang Shen ; Hui Qian ; Tengfei Zhou ; Tongzhou Mu
【Abstract】: Variance Reducing (VR) stochastic methods are fast-converging alternatives to the classical Stochastic Gradient Descent (SGD) for solving large-scale regularized finite sum problems, especially when a highly accurate solution is required. One critical step in VR is the function sampling. State-of-the-art VR algorithms such as SVRG and SAGA, employ either Uniform Probability (UP) or Importance Probability (IP), which is deficient in reducing the variance and hence leads to suboptimal convergence rate. In this paper, we propose a novel sampling scheme that explicitly computes some Adaptive Probability (AP) at each iteration. Analysis shows that, equipped with AP, both SVRG and SAGA yield provably better convergence rate than the ones with UP or IP, which is confirmed in experiments. Additionally, to cut down the per iteration computation load, an efficient variant is proposed by utilizing AP periodically, whose performance is empirically validated.
【Keywords】:
【Paper Link】 【Pages】:1997-2003
【Authors】: Lei Shi ; Yi-Dong Shen
【Abstract】: Convex Transductive Experimental Design (CTED) is one of the most representative active learning methods. It utilizes a data reconstruction framework to select informative samples for manual annotation. However, we observe that CTED cannot well handle the diversity of selected samples and hence the set of selected samples may contain mutually similar samples which convey similar or overlapped information. This is definitely undesired. Given limited budget for data labeling, it is desired to select informative samples with complementary information, i.e., similar samples are excluded. To this end, we proposes Diversified CTED by seamlessly incorporating a novel and effective diversity regularizer into CTED, ensuring the selected samples are diverse. The involvement of the diversity regularizer leads the optimization problem hard to solve. We derive an effective algorithm to solve an equivalent problem which is easier to optimize. Extensive experimental results on several benchmark data sets demonstrate that Diversified CTED significantly improves CTED and consistently outperforms the state-of-the-art methods, verifying the effectiveness and advantages of incorporating the proposed diversity regularizer into CTED.
【Keywords】:
【Paper Link】 【Pages】:2004-2010
【Authors】: Weiwei Shi ; Yihong Gong ; Jinjun Wang
【Abstract】: In this paper, we propose a novel method to improve object recognition accuracies of convolutional neural networks (CNNs) by embedding the proposed Min-Max objective into a high layer of the models during the training process. The Min-Max objective explicitly enforces the learned object feature maps to have the minimum compactness for each object manifold and the maximum margin between different object manifolds. The Min-Max objective can be universally applied to different CNN models with negligible additional computation cost. Experiments with shallow and deep models on four benchmark datasets including CIFAR-10, CIFAR-100, SVHN and MNIST demonstrate that CNN models trained with the Min-Max objective achieve remarkable performance improvements compared to the corresponding baseline models.
【Keywords】:
【Paper Link】 【Pages】:2011-2017
【Authors】: Harold Soh
【Abstract】: Embeddings or vector representations of objects have been used with remarkable success in various machine learning and AI tasks — from dimensionality reduction and data visualization, to vision and natural language processing. In this work, we seek probabilistic embeddings that faithfully represent observed relationships between objects (e.g., physical distances, preferences). We derive a novel variational Bayesian variant of multidimensional scaling that (i) provides a posterior distribution over latent points without computationally-heavy Markov chain Monte Carlo (MCMC) sampling, and (ii) can leverage existing side information using sparse Gaussian processes (GPs) to learn a nonlinear mapping to the embedding. By partitioning entities, our method naturally handles incomplete side information from multiple domains, e.g., in product recommendation where ratings are available, but not all users and items have associated profiles. Furthermore, the derived approximate bounds can be used to discover the intrinsic dimensionality of the data and limit embedding complexity. We demonstrate the effectiveness of our methods empirically on three synthetic problems and on the real-world tasks of political unfolding analysis and multi-sensor localization.
【Keywords】:
【Paper Link】 【Pages】:2018-2024
【Authors】: Dongjin Song ; Wei Liu ; David A. Meyer
【Abstract】: Binary coding techniques, which compress originally high-dimensional data samples into short binary codes, are becoming increasingly popular due to their efficiency for information retrieval. Leveraging supervised information can dramatically enhance the coding quality, and hence improve search performance. There are few methods, however, that efficiently learn coding functions that optimize the precision at the top of the Hamming distance ranking list while approximately preserving the geometric relationships between database examples. In this paper, we propose a novel supervised binary coding approach, namely Fast Structural Binary Coding (FSBC), to optimize the precision at the top of a Hamming distance ranking list and ensure that similar images can be returned as a whole. The key idea is to train disciplined coding functions by optimizing a lower bound of the area under the ROC (Receiver Operating Characteristic) curve (AUC) and penalize this objective so that the geometric relationships between database examples in the original Euclidean space are approximately preserved in the Hamming space. To find such a coding function, we relax the original discrete optimization objective with a continuous surrogate, and then derive a stochastic gradient descent method to optimize the surrogate objective efficiently. Empirical studies based upon two image datasets demonstrate that the proposed binary coding approaches achieve superior image search performance to the states-of-the-art.
【Keywords】:
【Paper Link】 【Pages】:2025-2031
【Authors】: Young Chol Song ; Iftekhar Naim ; Abdullah Al Mamun ; Kaustubh Kulkarni ; Parag Singla ; Jiebo Luo ; Daniel Gildea ; Henry A. Kautz
【Abstract】: Advances in video technology and data storage have made large scale video data collections of complex activities readily accessible. An increasingly popular approach for automatically inferring the details of a video is to associate the spatio-temporal segments in a video with its natural language descriptions. Most algorithms for connecting natural language with video rely on pre-aligned supervised training data. Recently, several models have been shown to be effective for unsupervised alignment of objects in video with language. However, it remains difficult to generate good spatio-temporal video segments for actions that align well with language. This paper presents a framework that extracts higher level representations of low-level action features through hyperfeature coding from video and aligns them with language. We propose a two-step process that creates a high-level action feature codebook with temporally consistent motions, and then applies an unsupervised alignment algorithm over the action codewords and verbs in the language to identify individual activities. We show an improvement over previous alignment models of objects and nouns on videos of biological experiments, and also evaluate our system on a larger scale collection of videos involving kitchen activities.
【Keywords】:
【Paper Link】 【Pages】:2032-2038
【Authors】: Ivan Stojkovic ; Vladisav Jelisavcic ; Veljko Milutinovic ; Zoran Obradovic
【Abstract】: Graphical models, as applied to multi-target prediction problems, commonly utilize interaction terms to impose structure among the output variables. Often, such structure is based on the assumption that related outputs need to be similar and interaction terms that force them to be closer are adopted. Here we relax that assumption and propose a feature that is based on distance and can adapt to ensure that variables have smaller or larger difference in values. We utilized a Gaussian Conditional Random Field model, where we have extended its originally proposed interaction potential to include a distance term. The extended model is compared to the baseline in various structured regression setups. An increase in predictive accuracy was observed on both synthetic examples and real-world applications, including challenging tasks from climate and healthcare domains.
【Keywords】:
【Paper Link】 【Pages】:2039-2045
【Authors】: Sanatan Sukhija ; Narayanan Chatapuram Krishnan ; Gurkanwal Singh
【Abstract】: Heterogeneity of features and lack of correspondence between data points of different domains are the two primary challenges while performing feature transfer. In this paper, we present a novel supervised domain adaptation algorithm (SHDA-RF) that learns the mapping between heterogeneous features of different dimensions. Our algorithm uses the shared label distributions present across the domains as pivots for learning a sparse feature transformation. The shared label distributions and the relationship between the feature spaces and the label distributions are estimated in a supervised manner using random forests. We conduct extensive experiments on three diverse datasets of varying dimensions and sparsity to verify the superiority of the proposed approach over other baseline and state of the art transfer approaches.
【Keywords】:
【Paper Link】 【Pages】:2046-2052
【Authors】: Jun Suzuki ; Masaaki Nagata
【Abstract】:
The word embedding vectors obtained from neural word embedding methods, such as vLBL models and SkipGram, have become an important fundamental resource for tackling a wide variety of tasks in the artificial intelligence field. This paper focuses on the fact that the model size of high-quality embedding vectors is relatively large, i.e., more than 1GB. We propose a learning framework that can provide a set of compact' embedding vectors for the purpose of enhancing
usability' in actual applications. Our proposed method incorporates parameter sharing constraints into the optimization problem. These additional constraints force the embedding vectors to share parameter values, which significantly shrinks model size. We investigate the trade-off between quality and model size of embedding vectors for several linguistic benchmark datasets, and show that our method can significantly reduce the model size while maintaining the task performance of conventional methods.
【Keywords】:
【Paper Link】 【Pages】:2053-2059
【Authors】: Hao Tang ; Hong Liu
【Abstract】: Feature-to-feature matching is the key issue in the Bag-of-Features model.The baseline approach employs a coarse feature-to-feature matching, namely, two descriptors are assumed to match if they are assigned the same quantization index.However, this Hard Assignment strategy usually incurs undesirable low precision.To fix it, Multiple Assignment and Soft Assignment are proposed.These two methods reduce the quantization error to some extent, but there are still a lot of room for improvement.To further improve retrieval precision, in this paper, we propose a novel feature matching strategy, called local-restricted Soft Assignment (lrSA), in which a new feature matching function is introduced.The lrSA strategy is evaluated through extensive experiments on five benchmark datasets.Experiments show that the results exceed the retrieval performance of current quantization methods on these datasets.Combined with post-processing steps, we have achieved competitive results compared with the state-of-the-art methods.Overall, our strategy shows notable benefit for retrieval with large vocabularies and dataset size.
【Keywords】:
【Paper Link】 【Pages】:2060-2066
【Authors】: Joseph G. Taylor ; Viktoriia Sharmanska ; Kristian Kersting ; David Weir ; Novi Quadrianto
【Abstract】: Feature selection has been studied in machine learning and data mining for many years, and is a valuable way to improve classification accuracy while reducing model complexity. Two main classes of feature selection methods — filter and wrapper — discard those features which are not selected, and do not consider them in the predictive model. We propose that these unselected features may instead be used as an additional source of information at train time. We describe a strategy called Learning using Unselected Features (LUFe) that allows selected and unselected features to serve different functions in classification. In this framework, selected features are used directly to set the decision boundary, and unselected features are utilised in a secondary role, with no additional cost at test time. Our empirical results on 49 textual datasets show that LUFe can improve classification performance in comparison with standard wrapper and filter feature selection.
【Keywords】:
【Paper Link】 【Pages】:2067-2073
【Authors】: Stefano Teso ; Andrea Passerini ; Paolo Viappiani
【Abstract】: In this paper we propose an approach to preference elicitation that is suitable to large configuration spaces beyond the reach of existing state-of-the-art approaches. Our setwise max-margin method can be viewed as a generalization of max-margin learning to sets, and can produce a set of diverse items that can be used to ask informative queries to the user. Moreover, the approach can encourage sparsity in the parameter space, in order to favor the assessment of utility towards combinations of weights that concentrate on just few features. We present a mixed integer linear programming formulation and show how our approach compares favourably with Bayesian preference elicitation alternatives and easily scales to realistic datasets.
【Keywords】:
【Paper Link】 【Pages】:2074-2081
【Authors】: Arun Venkatraman ; Wen Sun ; Martial Hebert ; Byron Boots ; J. Andrew Bagnell
【Abstract】: Data-driven approaches for learning dynamic models for Bayesian filtering often try to maximize the data likelihood given parametric forms for the transition and observation models. However, this objective is usually nonconvex in the parametrization and can only be locally optimized. Furthermore, learning algorithms typically do not provide performance guarantees on the desired Bayesian filtering task. In this work, we propose using inference machines to directly optimize the filtering performance. Our procedure is capable of learning partially-observable systems when the state space is either unknown or known in advance. To accomplish this, we adapt PREDICTIVE STATE INFERENCE MACHINES (PSIMs) by introducing the concept of hints, which incorporate prior knowledge of the state space to accompany the predictive state representation. This allows PSIM to be applied to the larger class of filtering problems which require prediction of a specific parameter or partial component of state. Our PSIM+HINTS adaptation enjoys theoretical advantages similar to the original PSIM algorithm, and we showcase its performance on a variety of robotics filtering problems.
【Keywords】:
【Paper Link】 【Pages】:2082-2088
【Authors】: Aäron Verachtert ; Hendrik Blockeel ; Jesse Davis
【Abstract】: Energy efficiency is a concern for any software running on mobile devices. As such software employs machine-learned models to make predictions, this motivates research on efficiently executable models. In this paper, we propose a variant of the widely used Naive Bayes (NB) learner that yields a more efficient predictive model. In contrast to standard NB, where the learned model inspects all features to come to a decision, or NB with feature selection, where the model uses a fixed subset of the features, our model dynamically determines, on a case-by-case basis, when to stop inspecting features. We show that our approach is often much more efficient than the current state of the art, without loss of accuracy.
【Keywords】:
【Paper Link】 【Pages】:2089-2096
【Authors】: Ngo Anh Vien ; Peter Englert ; Marc Toussaint
【Abstract】: Modeling policies in reproducing kernel Hilbert space (RKHS) renders policy gradient reinforcement learning algorithms non-parametric. As a result, the policies become very flexible and have a rich representational potential without a pre-defined set of features. However, their performances might be either non-covariant under re-parameterization of the chosen kernel, or very sensitive to step-size selection. In this paper, we propose to use a general framework to derive a new RKHS policy search technique. The new derivation leads to both a natural RKHS actor-critic algorithm and a RKHS expectation maximization (EM) policy search algorithm. Further,we show that kernelization enables us to learn in partially observable (POMDP) tasks which is considered daunting for parametric approaches. Via sparsification, a small set of "support vectors" representing the history is shown to be effectively discovered.For evaluations, we use three simulated (PO)MDP reinforcement learning tasks, and a simulated PR2's robotic manipulation task. The results demonstrate the effectiveness of the new RKHS policy search framework in comparison to plain RKHS actor-critic, episodic natural actor-critic, plain actor-critic, and PoWER approaches.
【Keywords】:
【Paper Link】 【Pages】:2097-2103
【Authors】: Boyu Wang ; Joelle Pineau
【Abstract】: While multitask learning has been extensively studied, most existing methods rely on linear models (e.g. linear regression, logistic regression), which may fail in dealing with more general (nonlinear) problems. In this paper, we present a new approach that combines dictionary learning with gradient boosting to achieve multitask learning with general (nonlinear) basis functions. Specifically, for each task we learn a sparse representation in a nonlinear dictionary that is shared across the set of tasks. Each atom of the dictionary is a nonlinear feature mapping of the original input space, learned in function space by gradient boosting. The resulting model is a hierarchical ensemble where the top layer of the hierarchy is the task-specific sparse coefficients and the bottom layer is the boosted models common to all tasks. The proposed method takes the advantages of both dictionary learning and boosting for multitask learning: knowledge across tasks can be shared via the dictionary, and flexibility and generalization performance are guaranteed by boosting. More important, this general framework can be used to adapt any learning algorithm to (nonlinear) multitask learning. Experimental results on both synthetic and benchmark real-world datasets confirm the effectiveness of the proposed approach for multitask learning.
【Keywords】:
【Paper Link】 【Pages】:2104-2110
【Authors】: De Wang ; Feiping Nie ; Heng Huang
【Abstract】: Human action recognition is important in improving human life in various aspects. However, the outliers and noise in data often bother the clustering tasks. Therefore, there is a great need for the robust data clustering techniques. Nonnegative matrix factorization (NMF) and Nonnegative Matrix Tri-Factorization (NMTF) methods have been widely researched these years and applied to many data clustering applications. With the presence of outliers, most previous NMF/NMTF models fail to achieve the optimal clustering performance. To address this challenge, in this paper, we propose three new NMF and NMTF models which are robust to outliers. Efficient algorithms are derived, which converge much faster than previous NMF methods and as fast as K-means algorithm, and scalable to large-scale data sets. Experimental results on both synthetic and real world data sets show that our methods outperform other NMF and NMTF methods in most cases, and in the meanwhile, take much less computational time.
【Keywords】:
【Paper Link】 【Pages】:2111-2117
【Authors】: Lu Wang ; Zhi-Hua Zhou
【Abstract】: Crowdsourcing is widely adopted in many domains as a popular paradigm to outsource work to individuals. In the machine learning community, crowdsourcing is commonly used as a cost-saving way to collect labels for training data. While a lot of effort has been spent on developing methods for inferring labels from a crowd, few work concentrates on the theoretical foundation of crowdsourcing learning. In this paper, we theoretically study the cost-saving effect of crowdsourcing learning, and present an upper bound for the minimally-sufficient number of crowd labels for effective crowdsourcing learning. Our results provide an understanding about how to allocate crowd labels efficiently, and are verified empirically.
【Keywords】:
【Paper Link】 【Pages】:2118-2124
【Authors】: Shuo Wang ; Leandro L. Minku ; Xin Yao
【Abstract】: Online class imbalance learning deals with data streams having very skewed class distributions in a timely fashion. Although a few methods have been proposed to handle such problems, most of them focus on two-class cases. Multi-class imbalance imposes additional challenges in learning. This paper studies the combined challenges posed by multi-class imbalance and online learning, and aims at a more effective and adaptive solution. First, we introduce two resampling-based ensemble methods, called MOOB and MUOB, which can process multi-class data directly and strictly online with an adaptive sampling rate. Then, we look into the impact of multi-minority and multi-majority cases on MOOB and MUOB in comparison to other methods under stationary and dynamic scenarios. Both multi-minority and multi-majority make a negative impact. MOOB shows the best and most stable G-mean in most stationary and dynamic cases.
【Keywords】:
【Paper Link】 【Pages】:2125-2131
【Authors】: Shuyang Wang ; Zhengming Ding ; Yun Fu
【Abstract】: In cross-domain learning, there is a more challenging problem that the domain divergence involves more than one dominant factors, e.g., different view-points, various resolutions and changing illuminations. Fortunately, an intermediate domain could often be found to build a bridge across them to facilitate the learning problem. In this paper, we propose a Coupled Marginalized Denoising Auto-encoders framework to address the cross-domain problem. Specifically, we design two marginalized denoising auto-encoders, one for the target and the other for source as well as the intermediate one. To better couple the two denoising auto-encoders learning, we incorporate a feature mapping, which tends to transfer knowledge between the intermediate domain and the target one. Furthermore, the maximum margin criterion, e.g., intra-class compactness and inter-class penalty, on the output layer is imposed to seek more discriminative features across different domains. Extensive experiments on two tasks have demonstrated the superiority of our method over the state-of-the-art methods.
【Keywords】:
【Paper Link】 【Pages】:2132-2138
【Authors】: William Yang Wang ; William W. Cohen
【Abstract】: Many complex reasoning tasks in Artificial Intelligence (including relation extraction, knowledge base completion, and information integration) can be formulated as inference problems using a probabilistic first-order logic. However, due to the discrete nature of logical facts and predicates, it is challenging to generalize symbolic representations and represent first-order logic formulas in probabilistic relational models. In this work, we take a rather radical approach: we aim at learning continuous low-dimensional embeddings for first-order logic from scratch. In particular, we first consider a structural gradient based structure learning approach to generate plausible inference formulas from facts; then, we build grounded proof graphs using background facts, training examples, and these inference formulas. To learn embeddings for formulas, we map the training examples into the rows of a binary matrix, and inference formulas into the columns. Using a scalable matrix factorization approach, we then learn the latent continuous representations of examples and logical formulas via a low-rank approximation method. In experiments, we demonstrate the effectiveness of reasoning with first-order logic embeddings by comparing with several state-of-the-art baselines on two datasets in the task of knowledge base completion.
【Keywords】:
【Paper Link】 【Pages】:2139-2145
【Authors】: Xin Wang ; Congfu Xu ; Yunhui Guo ; Hui Qian
【Abstract】: To learn users' preference, their feedback information is commonly modeled as scalars and integrated into matrix factorization (MF) based algorithms. Based on MF techniques, the preference degree is computed by the product of user and item vectors, which is also represented by scalars. On the contrary, in this paper, we express users' feedback as constrained vectors, and call the idea constrained preference embedding (CPE); it means that we regard users, items and all users' behavior as vectors. We find that this viewpoint is more flexible and powerful than traditional MF for item recommendation. For example, by the proposed assumption, users' heterogeneous actions can be coherently mined because all entities and actions can be transferred to a space of the same dimension. In addition, CPE is able to model the feedback of uncertain preference degree. To test our assumption, we propose two models called CPE-s and CPE-ps based on CPE for item recommendation, and show that the popular pair-wise ranking model BPR-MF can be deduced by some restrictions and variations on CPE-s. In the experiments, we will test CPE and the proposed algorithms, and prove their effectiveness.
【Keywords】:
【Paper Link】 【Pages】:2146-2152
【Authors】: Xuezhi Wang ; Junier B. Oliva ; Jeff G. Schneider ; Barnabás Póczos
【Abstract】: Multi-task learning attempts to simultaneously leverage data from multiple domains in order to estimate related functions on each domain. For example, a special case of multi-task learning, transfer learning, is often employed when one has a good estimate of a function on a source domain, but is unable to estimate a related function well on a target domain using only target data. Multi-task/transfer learning problems are usually solved by imposing some kind of "smooth" relationship among/between tasks. In this paper, we study how different smoothness assumptions on task relations affect the upper bounds of algorithms proposed for these problems under different settings. For general multi-task learning, we study a family of algorithms which utilize a reweighting matrix on task weights to capture the smooth relationship among tasks, which has many instantiations in existing literature. Furthermore, for multi-task learning in a transfer learning framework, we study the recently proposed algorithms for the "model shift", where the conditional distribution $P(Y|X)$ is allowed to change across tasks but the change is assumed to be smooth. In addition, we illustrate our results with experiments on both simulated and real data.
【Keywords】:
【Paper Link】 【Pages】:2153-2159
【Authors】: Yang Wang ; Wenjie Zhang ; Lin Wu ; Xuemin Lin ; Meng Fang ; Shirui Pan
【Abstract】: Multi-view spectral clustering, which aims at yielding an agreement or consensus data objects grouping across multi-views with their graph laplacian matrices, is a fundamental clustering problem. Among the existing methods, Low-Rank Representation (LRR) based method is quite superior in terms of its effectiveness, intuitiveness and robustness to noise corruptions. However, it aggressively tries to learn a common low-dimensional subspace for multi-view data, while inattentively ignoring the local manifold structure in each view, which is critically important to the spectral clustering; worse still, the low-rank minimization is enforced to achieve the data correlation consensus among all views, failing to flexibly preserve the local manifold structure for each view. In this paper, 1) we propose a multi-graph laplacian regularized LRR with each graph laplacian corresponding to one view to characterize its local manifold structure. 2) Instead of directly enforcing the low-rank minimization among all views for correlation consensus, we separately impose low-rank constraint on each view, coupled with a mutual structural consensus constraint, where it is able to not only well preserve the local manifold structure but also serve as a constraint for that from other views, which iteratively makes the views more agreeable. Extensive experiments on real-world multi-view data sets demonstrate its superiority.
【Keywords】:
【Paper Link】 【Pages】:2160-2166
【Authors】: Yi Wang ; Bin Li ; Xuhui Fan ; Yang Wang ; Fang Chen
【Abstract】: The Mondrian process (MP) produces hierarchical partitions on a product space as a kd-tree, which can be served as a flexible yet parsimonious partition prior for relational modeling. Due to the recursive generation of partitions and varying dimensionality of the partition state space, the inference procedure for the MP relational modeling is extremely difficult. The prevalent inference method reversible-jump MCMC for this problem requires a number of unnecessary retrospective steps to transit from one partition state to a very similar one and it is prone to fall into a local optimum. In this paper, we attempt to circumvent these drawbacks by proposing an alternative method for inferring the MP partition structure. Based on the observation that similar cutting rate measures on the partition space lead to similar partition layouts, we propose to impose a nonhomogeneous cutting rate measure on the partition space to control the layouts of the generated partitions — the original MCMC sampling problem is thus transformed into a Bayesian global optimization problem. The empirical tests demonstrate that Bayesian optimization is able to find better partition structures than MCMC sampling with the same number of partition structure proposals.
【Keywords】:
【Paper Link】 【Pages】:2167-2173
【Authors】: Yisen Wang ; Qingtao Tang ; Shu-Tao Xia ; Jia Wu ; Xingquan Zhu
【Abstract】: Random forests are one type of the most effective ensemble learning methods. In spite of their sound empirical performance, the study on their theoretical properties has been left far behind. Recently, several random forests variants with nice theoretical basis have been proposed, but they all suffer from poor empirical performance. In this paper, we propose a Bernoulli random forests model (BRF), which intends to close the gap between the theoretical consistency and the empirical soundness of random forests classification. Compared to Breiman's original random forests, BRF makes two simplifications in tree construction by using two independent Bernoulli distributions. The first Bernoulli distribution is used to control the selection of candidate attributes for each node of the tree, and the second one controls the splitting point used by each node. As a result, BRF enjoys proved theoretical consistency, so its accuracy will converge to optimum (i.e., the Bayes risk) as the training data grow infinitely large. Empirically, BRF demonstrates the best performance among all theoretical random forests, and is very comparable to Breiman's original random forests (which do not have the proved consistency yet). The theoretical and experimental studies advance the research one step further towards closing the gap between the theory and the practical performance of random forests classification.
【Keywords】:
【Paper Link】 【Pages】:2174-2180
【Authors】: Zhangyang Wang ; Yingzhen Yang ; Shiyu Chang ; Qing Ling ; Thomas S. Huang
【Abstract】: We investigate the L-infinity constrained representation which demonstrates robustness to quantization errors, utilizing the tool of deep learning. Based on the Alternating Direction Method of Multipliers (ADMM), we formulate the original convex minimization problem as a feed-forward neural network, named Deep L-infinity Encoder, by introducing the novel Bounded Linear Unit (BLU) neuron and modeling the Lagrange multipliers as network biases. Such a structural prior acts as an effective network regularization, and facilitates the model initialization. We then investigate the effective use of the proposed model in the application of hashing, by coupling the proposed encoders under a supervised pairwise loss, to develop a Deep Siamese L-infinity Network, which can be optimized from end to end. Extensive experiments demonstrate the impressive performances of the proposed model. We also provide an in-depth analysis of its behaviors against the competitors.
【Keywords】:
【Paper Link】 【Pages】:2181-2188
【Authors】: Zhe Wang ; Ling-Yu Duan ; Junsong Yuan ; Tiejun Huang ; Wen Gao
【Abstract】: We present a novel approach called Minimal Reconstruction Bias Hashing (MRH) to learn similarity preserving binary codes that jointly optimize both projection and quantization stages. Our work tackles an important problem of how to elegantly connect optimizing projection with optimizing quantization, and to maximize the complementary effects of two stages. Distinct from previous works, MRH can adaptively adjust the projection dimensionality to balance the information loss between projection and quantization. It is formulated as a problem of minimizing reconstruction bias of compressed signals. Extensive experiment results have shown the proposed MRH significantly outperforms a variety of state-of-the-art methods over several widely used benchmarks.
【Keywords】:
【Paper Link】 【Pages】:2189-2195
【Authors】: Pengfei Wei ; Yiping Ke ; Chi Keong Goh
【Abstract】: Deep feature learning has recently emerged with demonstrated effectiveness in domain adaptation. In this paper, we propose a Deep Nonlinear Feature Coding framework (DNFC) for unsupervised domain adaptation. DNFC builds on the marginalized stacked denoising autoencoder (mSDA) to extract rich deep features. We introduce two new elements to mSDA: domain divergence minimization by Maximum Mean Discrepancy (MMD), and nonlinear coding by kernelization. These two elements are essential for domain adaptation as they ensure the extracted deep features to have a small distribution discrepancy and encode data nonlinearity. The effectiveness of DNFC is verified by extensive experiments on benchmark datasets. Specifically, DNFC attains much higher prediction accuracy than state-of-the-art domain adaptation methods. Compared to its basis mSDA, DNFC is able to achieve remarkable prediction improvement and meanwhile converges much faster with a small number of stacked layers.
【Keywords】:
【Paper Link】 【Pages】:2196-2202
【Authors】: Felix Weninger ; Fabien Ringeval ; Erik Marchi ; Björn W. Schuller
【Abstract】: Continuous dimensional emotion recognition from audio is a sequential regression problem, where the goal is to maximize correlation between sequences of regression outputs and continuous-valued emotion contours, while minimizing the average deviation. As in other domains, deep neural networks trained on simple acoustic features achieve good performance on this task. Yet, the usual squared error objective functions for neural network training do not fully take into account the above-named goal. Hence, in this paper we introduce a technique for the discriminative training of deep neural networks using the concordance correlation coefficient as cost function, which unites both correlation and mean squared error in a single differentiable function. Results on the MediaEval 2013 and AV+EC 2015 Challenge data sets show that the proposed method can significantly improve the evaluation criteria compared to standard mean squared error training, both in the music and speech domains.
【Keywords】:
【Paper Link】 【Pages】:2203-2209
【Authors】: Nic Wilson ; Mojtaba Montazery
【Abstract】: One approach to preference learning, based on linear support vector machines, involves choosing a weight vector whose associated hyperplane has maximum margin with respect to an input set of preference vectors, and using this to compare feature vectors. However, as is well known, the result can be sensitive to how each feature is scaled, so that rescaling can lead to an essentially different vector. This gives rise to a set of possible weight vectors - which we call the rescale-optimal ones - considering all possible rescalings. From this set one can define a more cautious preference relation, in which one vector is preferred to another if it is preferred for all rescale-optimal weight vectors. In this paper, we analyse which vectors are rescale-optimal, and when there is a unique rescale-optimal vector, and we consider how to compute the induced preference relation.
【Keywords】:
【Paper Link】 【Pages】:2210-2216
【Authors】: Yingce Xia ; Tao Qin ; Weidong Ma ; Nenghai Yu ; Tie-Yan Liu
【Abstract】: We study the multi-play budgeted multi-armed bandit (MP-BMAB) problem, in which pulling an arm receives both a random reward and a random cost, and a player pulls L( ≥ 1) arms at each round. The player targets at maximizing her total expected reward under a budget constraint B for the pulling costs. We present a multiple ratio confidence bound policy: At each round, we first calculate a truncated upper (lower) confidence bound for the expected reward (cost) of each arm, and then pull the L arms with the maximum ratio of the sum of the upper confidence bounds of rewards to the sum of the lower confidence bounds of costs. We design 0-1 integer linear fractional programming oracle that can pick such the L arms within polynomial time. We prove that the regret of our policy is sublinear in general and is log-linear for certain parameter settings. We further consider two special cases of MP-BMABs: (1) We derive a lower bound for any consistent policy for MP-BMABs with Bernoulli reward and cost distributions. (2) We show that the proposed policy can also solve conventional budgeted MAB problem (a special case of MP-BMABs with L = 1) and provides better theoretical results than existing UCB-based pulling policies.
【Keywords】:
【Paper Link】 【Pages】:2217-2223
【Authors】: Liping Xie ; Dacheng Tao ; Haikun Wei
【Abstract】: Video-based facial expression recognition (FER)has recently received increased attention as a result of its widespread application. Many kinds of features have been proposed to represent different properties of facial expressions in videos. However the dimensionality of these features is usually high. In addition, due to the complexity of the information available in video sequences, using only one type of feature is often inadequate. How to effectively reduce the dimensionality and combine multi-view features thus becomes a challenging problem. In this paper, motivated by the recent success in exclusive feature selection, we first introduce exclusive group LASSO (EG-LASSO) to unsupervised dimension reduction (UDR). This leads to the proposed exclusive UDR (EUDR) framework, which allows arbitrary sparse structures on the feature space. To properly combine multiple kinds of features, we further extend EUDR to multi-view EUDR (MEUDR), where the structured sparsity is enforced at both intra- and inter-view levels. In addition, combination weights are learned for all views to allow them to contribute differently to the final consensus presentation. A reliable solution is then obtained. Experiments on two challenging video-based FER datasets demonstrate the effectiveness of the proposed method.
【Keywords】:
【Paper Link】 【Pages】:2224-2230
【Authors】: Jinglin Xu ; Junwei Han ; Kai Xiong ; Feiping Nie
【Abstract】: The partition-based clustering algorithms, like K-Means and fuzzy K-Means, are most widely and successfully used in data mining in the past decades. In this paper, we present a robust and sparse fuzzy K-Means clustering algorithm, an extension to the standard fuzzy K-Means algorithm by incorporating a robust function, rather than the square data fitting term, to handle outliers. More importantly, combined with the concept of sparseness, the new algorithm further introduces a penalty term to make the object-clusters membership of each sample have suitable sparseness. Experimental results on benchmark datasets demonstrate that the proposed algorithm not only can ensure the robustness of such soft clustering algorithm in real world applications, but also can avoid the performance degradation by considering the membership sparsity.
【Keywords】:
【Paper Link】 【Pages】:2231-2237
【Authors】: Yasunori Yamada ; Tetsuro Morimura
【Abstract】: Deep neural networks frequently require the careful tuning of model hyper parameters. Recent research has shown that automated early termination of underperformance runs can speed up hyper parameter searches. However, these studies have used only learning curve for predicting the eventual model performance. In this study, we propose using weight features extracted from network weights at an early stage of the learning process as explanation variables for predicting the eventual model performance. We conduct experiments on hyper parameter searches with various types of convolutional neural network architecture on three image datasets and apply the random forest method for predicting the eventual model performance. The results show that use of the weight features improves the predictive performance compared with use of the learning curve. In all three datasets, the most important feature for the prediction was related to weight changes in the last convolutional layers. Our findings demonstrate that using weight features can help construct prediction models with a smaller number of training samples and terminate underperformance runs at an earlier stage of the learning process of DNNs than the conventional use of learning curve, thus facilitating the speed-up of hyper parameter searches.
【Keywords】:
【Paper Link】 【Pages】:2238-2244
【Authors】: Rui Yan
【Abstract】: Part of the long lasting cultural heritage of humanity is the art of classical poems, which are created by fitting words into certain formats and representations. Automatic poetry composition by computers is considered as a challenging problem which requires high Artificial Intelligence assistance. This study attracts more and more attention in the research community. In this paper, we formulate the poetry composition task as a natural language generation problem using recurrent neural networks. Given user specified writing intents, the system generates a poem via sequential language modeling. Unlike the traditional one-pass generation for previous neural network models, poetry composition needs polishing to satisfy certain requirements. Hence, we propose a new generative model with a polishing schema, and output a refined poem composition. In this way, the poem is generated incrementally and iteratively by refining each line. We run experiments based on large datasets of 61,960 classic poems in Chinese. A comprehensive evaluation, using perplexity and BLEU measurements as well as human judgments, has demonstrated the effectiveness of our proposed approach.
【Keywords】:
【Paper Link】 【Pages】:2245-2251
【Authors】: Xiaoqiang Yan ; Yangdong Ye ; Xueying Qiu
【Abstract】: Recent researches have shown consensus clustering can enhance the accuracy of human action categorization models by combining multiple clusterings, which can be obtained from various types of local descriptors, such as HOG, HOF and MBH. However, consensus clustering yields final clustering without access to the underlying feature representations of the human action data, which always makes the final partition limited to the quality of existing basic clusterings. To solve this problem, we present a novel and effective Consensus Information Bottleneck (CIB) method for unsupervised human action categorization. CIB is capable of learning action categories from feature variable and auxiliary clusterings simultaneously. Specifically, by performing Maximization of Mutual Information (MMI), CIB maximally preserves the information between feature variable and existing auxiliary clusterings. Moreover, to solve MMI optimization, a sequential solution is proposed to update data partition. Extensive experiments on five realistic human action data sets show that CIB can consistently and significantly beat other state-of-the-art consensus and multi-view clustering methods.
【Keywords】:
【Paper Link】 【Pages】:2252-2258
【Authors】: Liang Yang ; Xiaochun Cao ; Dongxiao He ; Chuan Wang ; Xiao Wang ; Weixiong Zhang
【Abstract】: Identification of module or community structures is important for characterizing and understanding complex systems. While designed with different objectives, i.e., stochastic models for regeneration and modularity maximization models for discrimination, both these two types of model look for low-rank embedding to best represent and reconstruct network topology. However, the mapping through such embedding is linear, whereas real networks have various nonlinear features, making these models less effective in practice. Inspired by the strong representation power of deep neural networks, we propose a novel nonlinear reconstruction method by adopting deep neural networks for representation. We then extend the method to a semi-supervised community detection algorithm by incorporating pairwise constraints among graph nodes. Extensive experimental results on synthetic and real networks show that the new methods are effective, outperforming most state-of-the-art methods for community detection.
【Keywords】:
【Paper Link】 【Pages】:2259-2265
【Authors】: Xu Yang ; Xin Geng ; Deyu Zhou
【Abstract】: By observing that the faces at close ages are similar, some Label Distribution Learning (LDL) methods have been proposed to solve age estimation tasks that they treat age distributions as the training targets. However, these existent LDL methods are limited because they can hardly extract enough useful information from complex image features. In this paper, Sparsity Conditional Energy Label Distribution Learning (SCE-LDL) is proposed to solve this problem. In the proposed SCE-LDL, age distributions are used as the training targets and energy function is utilized to define the age distribution. By assigning a suitable energy function, SCE-LDL can learn distributed representations, which provides the model with strong expressiveness for capturing enough of the complexity of interest from image features. The sparsity constraints are also incorporated to ameliorate the model. Experiment results in two age datasets show remarkable advantages of the proposed SCE-LDL model over the previous proposed age estimation methods.
【Keywords】:
【Paper Link】 【Pages】:2266-2272
【Authors】: Xun Yang ; Meng Wang ; Luming Zhang ; Dacheng Tao
【Abstract】: Traditional metric learning methods usually make decisions based on a fixed threshold, which may result in a suboptimal metric when the inter-class and inner-class variations are complex. To address this issue, in this paper we propose an effective metric learning method by exploiting privileged information to relax the fixed threshold under the empirical risk minimization framework. Privileged information describes useful high-level semantic information that is only available during training. Our goal is to improve the performance by incorporating privileged information to design a locally adaptive decision function. We jointly learn two distance metrics by minimizing the empirical loss penalizing the difference between the distance in the original space and that in the privileged space. The distance in the privileged space functions as a locally adaptive decision threshold, which can guide the decision making like a teacher. We optimize the objective function using the Accelerated Proximal Gradient approach to obtain a global optimum solution. Experiment results show that by leveraging privileged information, our proposed method can achieve satisfactory performance.
【Keywords】:
【Paper Link】 【Pages】:2273-2279
【Authors】: Yang Yang ; Fumin Shen ; Zi Huang ; Heng Tao Shen
【Abstract】: Spectral clustering has been playing a vital role in various research areas. Most traditional spectral clustering algorithms comprise two independent stages (i.e., first learning continuous labels and then rounding the learned labels into discrete ones), which may lead to severe information loss and performance degradation. In this work, we study how to achieve discrete clustering as well as reliably generalize to unseen data. We propose a unified spectral clustering scheme which jointly learns discrete clustering labels and robust out-of-sample prediction functions. Specifically, we explicitly enforce a discrete transformation on the intermediate continuous labels, which leads to a tractable optimization problem with a discrete solution. Moreover, to further compensate the unreliability of the learned labels, we integrate an adaptive robust module with ℓ2,p loss to learn prediction function for unseen data. Extensive experiments conducted on various data sets have demonstrated the superiority of our proposal as compared to existing clustering approaches.
【Keywords】:
【Paper Link】 【Pages】:2280-2286
【Authors】: Yang Yang ; De-Chuan Zhan ; Yuan Jiang
【Abstract】: Complex objects are usually with multiple modal features. In multi-modal learning, modalities closely related to the target tasks are known as strong modalities. While collecting strong modalities of all instances is often expensive, and current multi-modal learning techniques hardly take the strong modal feature extraction expenses into consideration. On the other hand, active learning is proposed to reduce the labeling expenses by querying the ground truths for specific selected instances. In this paper, we propose a training strategy, ACQUEST (ACtive QUErying STrong modalities), which exploits strong modal information by actively querying the strong modal feature values of "selected" instances rather than their corresponding ground truths. In ACQUEST, only the informative instances are selected for strong modal feature acquisition. An inverse prediction technique is also proposed to make the ACQUEST a unified optimization form. Experiments on image datasets show that ACQUEST achieves better classification performance than conventional active learning and multi-modal learning methods with less feature acquisition costs and labeling expenses.
【Keywords】:
【Paper Link】 【Pages】:2287-2293
【Authors】: Zhilin Yang ; Jie Tang ; William W. Cohen
【Abstract】: We study the extent to which online social networks can be connected to open knowledge bases. The problem is referred to as learning social knowledge graphs.We propose a multi-modal Bayesian embedding model, GenVector, to learn latent topics that generate word and network embeddings.GenVector leverages large-scale unlabeled data with embeddings and represents data of two modalities — i.e., social network users and knowledge concepts — in a shared latent topic space.Experiments on three datasets show that the proposed method clearly outperforms state-of-the-art methods. We then deploy the method on AMiner, a large-scale online academic search system with a network of 38,049,189 researchers with a knowledge base with 35,415,011 concepts. Our method significantly decreases the error rate in an online A/B test with live users.
【Keywords】:
【Paper Link】 【Pages】:2294-2300
【Authors】: Quanming Yao ; James T. Kwok
【Abstract】: Learning of low-rank matrices is fundamental to many machine learning applications. A state-of-the-art algorithm is the rank-one matrix pursuit (R1MP). However, it can only be used in matrix completion problems with the square loss. In this paper, we develop a more flexible greedy algorithm for generalized low-rank models whose optimization objective can be smooth or nonsmooth, general convex or strongly convex. The proposed algorithm has low per-iteration time complexity and fast convergence rate.Experimental results show that it is much faster than the state-of-the-art,with comparable or even better prediction performance.
【Keywords】:
【Paper Link】 【Pages】:2301-2307
【Authors】: Qiaomin Ye ; Luo Luo ; Zhihua Zhang
【Abstract】: Approximate matrix multiplication (AMM) becomes increasingly popular because it makes matrix computation suitable for large-scale datasets. Most previous AMM methods are based on the idea of random selection or random projection. In this paper, we propose a deterministic algorithm FD-AMM for computing an approximation to the product of two given matrices. Moreover, the algorithm works in a streaming manner. In particular, our approach is inspired by a recently proposed matrix sketching algorithm called Frequent Directions (FD). FD-AMM has stronger error bound than both random selection and random projection algorithms with respect to the same space complexity. Our approach also leads to an algorithm for computing the Canonical Correlation Analysis (CCA) of two matrices exactly in a streaming way, which takes less space than the classical method. Experimental results validate the effectiveness of our method.
【Keywords】:
【Paper Link】 【Pages】:2308-2314
【Authors】: Pengcheng Yin ; Zhengdong Lu ; Hang Li ; Ben Kao
【Abstract】: We propose Neural Enquirer — a neural network architecture for answering natural language (NL) questions based on a knowledge base (KB) table. Unlike existing work on end-to-end training of semantic parsers, Neural Enquirer is fully "neuralized": it finds distributed representations of queries and KB tables, and executes queries through a series of neural network components called "executors". Executors model query operations and compute intermediate execution results in the form of table annotations at different levels. Neural Enquirer can be trained with gradient descent, with which the representations of queries and the KB table are jointly optimized with the query execution logic. The training can be done in an end-to-end fashion, and it can also be carried out with stronger guidance, e.g., step-by-step supervision for complex queries. Neural Enquirer is one step towards building neural network systems that can understand natural language in real-world tasks. As a proof-of-concept, we conduct experiments on a synthetic QA task, and demonstrate that the model can learn to execute reasonably complex NL queries on small-scale KB tables.
【Keywords】:
【Paper Link】 【Pages】:2315-2321
【Authors】: Yusen Zhan ; Haitham Bou-Ammar ; Matthew E. Taylor
【Abstract】: Policy advice is a transfer learning method where a student agent is able to learn faster via advice from a teacher. However, both this and other reinforcement learning transfer methods have little theoretical analysis. This paper formally defines a setting where multiple teacher agents can provide advice to a student and introduces an algorithm to leverage both autonomous exploration and teacher's advice. Our regret bounds justify the intuition that good teachers help while bad teachers hurt. Using our formalization, we are also able to quantify, for the first time, when negative transfer can occur within such a reinforcement learning setting.
【Keywords】:
【Paper Link】 【Pages】:2322-2328
【Authors】: Qin Zhang ; Jia Wu ; Hong Yang ; Yingjie Tian ; Chengqi Zhang
【Abstract】: In this paper we study the problem of learning discriminative features (segments), often referred to as shapelets [Ye and Keogh, 2009] of time series, from unlabeled time series data. Discovering shapelets for time series classification has been widely studied, where many search-based algorithms are proposed to efficiently scan and select segments from a pool of candidates. However, such types of search-based algorithms may incur high time cost when the segment candidate pool is large. Alternatively, a recent work [Grabocka et al., 2014] uses regression learning to directly learn, instead of searching for, shapelets from time series. Motivated by the above observations, we propose a new Unsupervised Shapelet Learning Model (USLM) to efficiently learn shapelets from unlabeled timeseries data. The corresponding learning function integrates the strengths of pseudo-class label, spectral analysis, shapelets regularization term and regularized least-squares to auto-learn shapelets, pseudo-class labels and classification boundaries simultaneously. A coordinate descent algorithm is used to iteratively solve the learning function. Experiments show that USLM outperforms search-based algorithms on real-world time series data.
【Keywords】:
【Paper Link】 【Pages】:2329-2335
【Authors】: Qing Zhang ; Houfeng Wang
【Abstract】: Real-world data are seldom unstructured, yet traditional Matrix Factorization (MF) models, as one of the most powerful collaborative filtering approaches, generally rely on this assumption to recover the low-rank structures for recommendation. However, few of them are able to explicitly consider structured constraint with the underlying low-rank assumption to model complex user interests. To solve this problem, we propose a unified MF framework with generalized Laplacian constraint for collaborative filtering. We investigate the connection between the recently proposed Laplacian constraint and the classical normalized cut problem, and make it possible to extend the original non-overlapping prior, to capture the overlapping case via learning the decomposed multi-facet graphs. Experiments on real-world datasets demonstrate the effectiveness of the proposed method.
【Keywords】:
【Paper Link】 【Pages】:2336-2342
【Authors】: Ruqi Zhang ; Zhiwu Lu
【Abstract】: Large-scale clustering has found wide applications in many fields and received much attention in recent years. However, most existing large-scale clustering methods can only achieve mediocre performance, because they are sensitive to the unavoidable presence of noise in the large-scale data. To address this challenging problem, we thus propose a large-scale sparse clustering (LSSC) algorithm. In this paper, we choose a two-step optimization strategy for large-scale sparse clustering: 1) k-means clustering over the large-scale data to obtain the initial clustering results; 2) clustering refinement over the initial results by developing a spare coding algorithm. To guarantee the scalability of the second step for large-scale data, we also utilize nonlinear approximation and dimension reduction techniques to speed up the sparse coding algorithm. Experimental results on both synthetic and real-world datasets demonstrate the promising performance of our LSSC algorithm.
【Keywords】:
【Paper Link】 【Pages】:2343-2349
【Authors】: Shizhou Zhang ; Yihong Gong ; Jinjun Wang
【Abstract】: In this paper, we choose to learn useful cues from object recognition mechanisms of the human visual cortex, and propose a DCNN performance improvement method without the need for increasing the network complexity. Inspired by the category-selective property of the neuron population in the IT layer of the human visual cortex, we enforce the neuron responses at the top DCNN layer to be category selective. To achieve this, we propose the Sparse Category-Selective Objective Function to modulate the neuron outputs of the top DCNN layer. The proposed method is generic and can be applied to any DCNN models. As experimental results show, when applying the proposed method to the "Quick" model and NIN models, image classification performances are remarkably improved on four widely used benchmark datasets: CIFAR-10, CIFAR-100, MNIST and SVHN, which demonstrate the effectiveness of the presented method.
【Keywords】:
【Paper Link】 【Pages】:2350-2356
【Authors】: Wei Zhang ; Suyog Gupta ; Xiangru Lian ; Ji Liu
【Abstract】: Deep neural networks have been shown to achieve state-of-the-art performance in several machine learning tasks. Stochastic Gradient Descent (SGD) is the preferred optimization algorithm for training these networks and asynchronous SGD (ASGD) has been widely adopted for accelerating the training of large-scale deep networks in a distributed computing environment. However, in practice it is quite challenging to tune the training hyperparameters (such as learning rate) when using ASGD so as achieve convergence and linear speedup, since the stability of the optimization algorithm is strongly influenced by the asynchronous nature of parameter updates. In this paper, we propose a variant of the ASGD algorithm in which the learning rate is modulated according to the gradient staleness and provide theoretical guarantees for convergence of this algorithm. Experimental verification is performed on commonly-used image classification benchmarks: CIFAR10 and Imagenet to demonstrate the superior effectiveness of the proposed approach, compared to SSGD (Synchronous SGD) and the conventional ASGD algorithm.
【Keywords】:
【Paper Link】 【Pages】:2357-2363
【Authors】: Xianchao Zhang ; Xiaotong Zhang ; Han Liu
【Abstract】: Multi-task clustering improves the clustering performance of each task by transferring knowledge across related tasks. Most existing multi-task clustering methods are based on the ideal assumption that the tasks are completely related. However, in many real applications, the tasks are usually partially related, and brute-force transfer may cause negative effect which degrades the clustering performance.In this paper, we propose a self-adapted multi-task clustering (SAMTC) method which can automatically identify and transfer reusable instances among the tasks, thus avoiding negative transfer. SAMTC begins with an initialization by performing single-task clustering on each task, then executes the following three steps: first, it finds the reusable instances by measuring related clusters with Jensen-Shannon divergence between each pair of tasks, and obtains a pair of possibly related subtasks;second, it estimates the relatedness between each pair of subtasks with kernel mean matching;third, it constructs the similarity matrix for each task by exploiting useful information from the other tasks through instance transfer, and adopts spectral clustering to get the final clustering result. Experimental results on several real data sets show the superiority of the proposed algorithm over traditional single-task clustering methods and existing multitask clustering methods.
【Keywords】:
【Paper Link】 【Pages】:2364-2370
【Authors】: Yizhe Zhang ; Ricardo Henao ; Chunyuan Li ; Lawrence Carin
【Abstract】: In dictionary learning for analysis of images, spatial correlation from extracted patches can be leveraged to improve characterization power. We propose a Bayesian framework for dictionary learning, with spatial location dependencies captured by imposing a multiplicative Gaussian process (GP) priors on the latent units representing binary activations. Data augmentation and Kronecker methods allow for efficient Markov chain Monte Carlo sampling. We further extend the model with Sigmoid Belief Networks (SBNs), linking the GPs to the top-layer latent binary units of the SBN, capturing inter-dictionary dependencies while also yielding computational savings. Applications to image denoising, inpainting and depth-information restoration demonstrate that the proposed model outperforms other leading Bayesian dictionary learning approaches.
【Keywords】:
【Paper Link】 【Pages】:2371-2377
【Authors】: Zemin Zhang ; Shuchin Aeron
【Abstract】: In this paper a new dictionary learning algorithm for multidimensional data is proposed. Unlike most conventional dictionary learning methods which are derived for dealing with vectors or matrices, our algorithm, named K-TSVD, learns a multidimensional dictionary directly via a novel algebraic approach for tensor factorization as proposed in [Braman, 2010; Kilmer et al., 2011; Kilmer and Martin, 2011]. Using this approach one can define a tensor-SVD and we propose to extend K-SVD algorithm used for 1-D data to a K-TSVD algorithm for handling 2-D and 3-D data. Our algorithm, based on the idea of sparse coding (using group-sparsity over multidimensional coefficient vectors), alternates between estimating a compact representation and dictionary learning. We analyze our K-TSVD algorithm and demonstrate its result on video completion and video/multispectral image denoising.
【Keywords】:
【Paper Link】 【Pages】:2378-2384
【Authors】: Feipeng Zhao ; Yuhong Guo
【Abstract】: Personalized top-N recommendation systems have great impact on many real world applications such as E-commerce platforms and social networks. Most existing methods produce personalized top-N recommendations by minimizing a specific uniform loss such as pairwise ranking loss or pointwise recovery loss. In this paper, we propose a novel personalized top-N recommendation approach that minimizes a combined heterogeneous loss based on linear self-recovery models. The heterogeneous loss integrates the strengths of both pairwise ranking loss and pointwise recovery loss to provide more informative recommendation predictions. We formulate the learning problem with heterogeneous loss as a constrained convex minimization problem and develop a projected stochastic gradient descent optimization algorithm to solve it. We evaluate the proposed approach on a set of personalized top-N recommendation tasks. The experimental results show the proposed approach outperforms a number of state-of-the-art methods on top-N recommendation.
【Keywords】:
【Paper Link】 【Pages】:2385-2391
【Authors】: Feipeng Zhao ; Min Xiao ; Yuhong Guo
【Abstract】: Recommender systems have been widely studied in the literature as they have real world impacts in many E-commerce platforms and social networks. Most previous systems are based on the user-item recommendation matrix, which contains users' history recommendation activities on items. In this paper, we propose a novel predictive collaborative filtering approach that exploits both the partially observed user-item recommendation matrix and the item-based side information to produce top-N recommender systems. The proposed approach automatically identifies the most interesting items for each user from his or her non-recommended item pool by aggregating over his or her recommended items via a low-rank coefficient matrix. Moreover, it also simultaneously builds linear regression models from the item-based side information such as item reviews to predict the item recommendation scores for the users. The proposed approach is formulated as a rank constrained joint minimization problem with integrated least squares losses, for which an efficient analytical solution can be derived. To evaluate the proposed learning technique, empirical evaluations on five recommendation tasks are conducted. The experimental results demonstrate the efficacy of the proposed approach comparing to the competing methods.
【Keywords】:
【Paper Link】 【Pages】:2392-2398
【Authors】: Handong Zhao ; Hongfu Liu ; Yun Fu
【Abstract】: Nowadays multi-modal visual data are much easier to access as the technology develops. Nevertheless, there is an underlying problem hidden behind the emerging multi-modality techniques: What if one/more modal data fail? Motivated by this question, we propose an unsupervised method which well handles the incomplete multi-modal data by transforming the original and incomplete data to a new and complete representation in a latent space. Different from the existing efforts that simply project data from each modality into a common subspace, a novel graph Laplacian term with a good probabilistic interpretation is proposed to couple the incomplete multi-modal samples. In such a way, a compact global structure over the entire heterogeneous data is well preserved, leading to a strong grouping discriminability. As a non-trivial contribution, we provide the optimization solution to the proposed model. In experiments, we extensively test our method and competitors on one synthetic data, two RGB-D video datasets and two image datasets. The superior results validate the benefits of the proposed method, especially when multi-modal data suffer from large incompleteness.
【Keywords】:
【Paper Link】 【Pages】:2399-2406
【Authors】: Feng Zheng ; Ling Shao
【Abstract】: In this paper, we propose to learn cross-view binary identities (CBI) for fast person re-identification. To achieve this, two sets of discriminative hash functions for two different views are learned by simultaneously minimising their distance in the Hamming space, and maximising the cross-covariance and margin. Thus, similar binary codes can be found for images of a same person captured at different views by embedding the images into the Hamming space. Therefore, person re-identification can be solved by efficiently computing and ranking the Hamming distances between the images. Extensive experiments are conducted on two public datasets and CBI produces comparable results as state-ofthe- art re-identification approaches but is at least 2200 times faster.
【Keywords】:
【Paper Link】 【Pages】:2407-2613
【Authors】: Shuai Zheng ; James T. Kwok
【Abstract】: The alternating direction method of multipliers (ADMM) is a powerful optimization solver in machine learning. Recently, stochastic ADMM has been integrated with variance reduction methods for stochastic gradient, leading to SAG-ADMM and SDCA-ADMM that have fast convergence rates and low iteration complexities. However, their space requirements can still be high. In this paper, we propose an integration of ADMM with the method of stochastic variance reduced gradient (SVRG). Unlike another recent integration attempt called SCAS-ADMM, the proposed algorithm retains the fast convergence benefits of SAG-ADMM and SDCA-ADMM, but is more advantageous in that its storage requirement is very low, even independent of the sample size $n$. Experimental results demonstrate that it is as fast as SAG-ADMM and SDCA-ADMM, much faster than SCAS-ADMM, and can be used on much bigger data sets.
【Keywords】:
【Paper Link】 【Pages】:2414-2420
【Authors】: Joey Tianyi Zhou ; Xinxing Xu ; Sinno Jialin Pan ; Ivor W. Tsang ; Zheng Qin ; Rick Siow Mong Goh
【Abstract】: Most existing learning to hash methods assume that there are sufficient data, either labeled or unlabeled, on the domain of interest (i.e., the target domain) for training. However, this assumption cannot be satisfied in some real-world applications. To address this data sparsity issue in hashing, inspired by transfer learning, we propose a new framework named Transfer Hashing with Privileged Information (THPI). Specifically, we extend the standard learning to hash method, Iterative Quantization (ITQ), in a transfer learning manner, namely ITQ+. In ITQ+, a new slack function is learned from auxiliary data to approximate the quantization error in ITQ. We developed an alternating optimization approach to solve the resultant optimization problem for ITQ+. We further extend ITQ+ to LapITQ+ by utilizing the geometry structure among the auxiliary data for learning more precise binary codes in the target domain. Extensive experiments on several benchmark datasets verify the effectiveness of our proposed approaches through comparisons with several state-of-the-art baselines.
【Keywords】:
【Paper Link】 【Pages】:2421-2427
【Authors】: Xingyi Zhou ; Qingfu Wan ; Wei Zhang ; Xiangyang Xue ; Yichen Wei
【Abstract】: Previous learning based hand pose estimation methods does not fully exploit the prior information in hand model geometry. Instead, they usually rely a separate model fitting step to generate valid hand poses. Such a post processing is inconvenient and sub-optimal. In this work, we propose a model based deep learning approach that adopts a forward kinematics based layer to ensure the geometric validity of estimated poses. For the first time, we show that embedding such a non-linear generative process in deep learning is feasible for hand pose estimation. Our approach is verified on challenging public datasets and achieves state-of-the-art performance.
【Keywords】:
【Paper Link】 【Pages】:2428-2434
【Authors】: Yang Zhou ; Haiping Lu
【Abstract】: As a classical subspace learning method, Probabilistic PCA (PPCA) has been extended to several bilinear variants for dealing with matrix observations. However, they are all based on the Tucker model, leading to a restricted subspace representation and the problem of rotational ambiguity. To address these problems, this paper proposes a bilinear PPCA method named as Probabilistic Rank-One Matrix Analysis (PROMA). PROMA is based on the CP model, which leads to a more flexible subspace representation and does not suffer from rotational ambiguity. For better generalization, concurrent regularization is introduced to regularize the whole matrix subspace, rather than column and row factors separately. Experiments on both synthetic and real-world data demonstrate the superiority of PROMA in subspace estimation and classification as well as the effectiveness of concurrent regularization in regularizing bilinear PPCAs.
【Keywords】:
【Paper Link】 【Pages】:2435-2441
【Authors】: Yao Zhou ; Jingrui He
【Abstract】: Nowadays, the rapid proliferation of data makes it possible to build complex models for many real applications. Such models, however, usually require large amount of labeled data, and the labeling process can be both expensive and tedious for domain experts. To address this problem, researchers have resorted to crowdsourcing to collect labels from non-experts with much less cost. The key challenge here is how to infer the true labels from the large number of noisy labels provided by non-experts. Different from most existing work on crowdsourcing, which ignore the structure information in the labeling data provided by non-experts, in this paper, we propose a novel structured approach based on tensor augmentation and completion. It uses tensor representation for the labeled data, augments it with a ground truth layer, and explores two methods to estimate the ground truth layer via low rank tensor completion. Experimental results on 6 real data sets demonstrate the superior performance of the proposed approach over state-of-the-art techniques.
【Keywords】:
【Paper Link】 【Pages】:2442-2448
【Authors】: Pengfei Zhu ; Lei Zhang ; Wangmeng Zuo ; Xiangchu Feng ; Qinghua Hu
【Abstract】: Almost all the existing representation based classifiers represent a query sample as a linear combination of training samples, and their time and memory cost will increase rapidly with the number of training samples. We investigate the representation based classification problem from a rather different perspective in this paper, that is, we learn how each feature (i.e., each element) of a sample can be represented by the features of itself. Such a self-representation property of sample features can be readily employed for pattern classification and a novel self-representation induced classifier (SRIC) is proposed. SRIC learns a self-representation matrix for each class. Given a query sample, its self-representation residual can be computed by each of the learned self-representation matrices, and classification can then be performed by comparing these residuals. In light of the principle of SRIC, a discriminative SRIC (DSRIC) method is developed. For each class, a discriminative self-representation matrix is trained to minimize the self-representation residual of this class while representing little the features of other classes. Experimental results on different pattern recognition tasks show that DSRIC achieves comparable or superior recognition rate to state-of-the-art representation based classifiers, however, it is much more efficient and needs much less storage space.
【Keywords】:
【Paper Link】 【Pages】:2449-2457
【Authors】: Xiaojin Zhu ; Ara Vartanian ; Manish Bansal ; Duy Nguyen ; Luke Brandl
【Abstract】: We introduce a new topological feature representation for point cloud objects. Specifically, we construct a Stochastic Multiresolution Persistent Homology (SMURPH) kernel which represents an object's persistent homology at different resolutions. Under the SMURPH kernel two objects are similar if they have similar number and sizes of "holes" at these resolutions. Our multiresolution kernel can capture both global topology and fine-grained topological texture in the data. Importantly, on large point clouds the SMURPH kernel is more computationally tractable compared to existing topological data analysis methods. We demonstrate SMURPH's potential for clustering and classification on several applications, including eye disease classification and human activity recognition.
【Keywords】:
【Paper Link】 【Pages】:2458-2464
【Authors】: Andrés Abeliuk ; Gerardo Berbeglia ; Felipe Maldonado ; Pascal Van Hentenryck
【Abstract】: We study dynamic trial-offer markets, in which participants first try a product and later decide whether to purchase it or not. In these markets, social influence and position biases have a greater effect on the decisions taken in the sampling stage than those in the buying stage. We consider a myopic policy that maximizes the market efficiency for each incoming participant, taking into account the inherent quality of products, position biases, and social influence. We prove that this myopic policy is optimal and predictable asymptotically.
【Keywords】:
【Paper Link】 【Pages】:2465-2471
【Authors】: Arpita Biswas ; Ragavendran Gopalakrishnan ; Partha Dutta
【Abstract】: With the increase in adoption of Electric Vehicles (EVs), proper utilization of the charging infrastructure is an emerging challenge for service providers. Overstaying of an EV after a charging event is a key contributor to low utilization. Since overstaying is easily detectable by monitoring the power drawn from the charger, managing this problem primarily involves designing an appropriate penalty during the overstaying period. Higher penalties do discourage overstaying; however, due to uncertainty in parking duration, less people would find such penalties acceptable, leading to decreased utilization (and revenue). To analyze this central tradeoff, we develop a novel framework that integrates models for realistic user behavior into queueing dynamics to locate the optimal penalty from the points of view of utilization and revenue, for different values of the external charging demand. Next, when the model parameters are unknown, we show how an online learning algorithm, such as UCB, can be adapted to learn the optimal penalty. Our experimental validation, based on charging data from London, shows that an appropriate penalty can increase both utilization and revenue while significantly reducing overstaying.
【Keywords】:
【Paper Link】 【Pages】:2472-2478
【Authors】: Abdelhamid Boudane ; Saïd Jabbour ; Lakhdar Sais ; Yakoub Salhi
【Abstract】: Discovering association rules from transaction databases is one of the most studied data mining task. Many effective techniques have been proposed over the years. All these algorithms share the same two steps methodology: frequent itemsets enumeration followed by effective association rules generation step. In this paper, we propose a new propositional satisfiability based approach to mine association rules in a single step. The task is modeled as a Boolean formula whose models correspond to the rules to be mined. To highlight the flexibility of our proposed framework, we also address two other variants, namely the closed and indirect association rules mining tasks. Experiments on many datasets show that on both closed and indirect association rules mining tasks, our declarative approach achieves better performance than the state-of-the-art specialized techniques.
【Keywords】:
【Paper Link】 【Pages】:2479-2485
【Authors】: Gong Cheng ; Weixi Zhu ; Ziwei Wang ; Jianghui Chen ; Yuzhong Qu
【Abstract】: Answering questions in a university's entrance examination like Gaokao in China challenges AI technology. As a preliminary attempt to take up this challenge, we focus on multiple-choice questions in Gaokao, and propose a three-stage approach that exploits and extends information retrieval techniques. Taking Wikipedia as the source of knowledge, our approach obtains knowledge relevant to a question by retrieving pages from Wikipedia via string matching and context-based disambiguation, and then ranks and filters pages using multiple strategies to draw critical evidence, based on which the truth of each option is assessed via relevance-based entailment. It achieves encouraging results on real-life questions in recent history tests, significantly outperforming baseline approaches.
【Keywords】:
【Paper Link】 【Pages】:2486-2492
【Authors】: Keith Clark ; Bernhard Hengst ; Maurice Pagnucco ; David Rajaratnam ; Peter Robinson ; Claude Sammut ; Michael Thielscher
【Abstract】: This paper establishes a framework that hierarchically integrates symbolic and sub-symbolic representations in an architecture for cognitive robotics. It is formalised abstractly as nodes in a hierarchy, with each node a sub-task that maintains its own belief-state and generates behaviour. An instantiation is developed for a real robot building towers of blocks, subject to human interference; this hierarchy uses a node with a concurrent multitasking teleo-reactive program, a node embedding a physics simulator to provide spatial knowledge, and nodes for sensor processing and robot control.
【Keywords】:
【Paper Link】 【Pages】:2493-2499
【Authors】: Dustin Dannenhauer ; Hector Muñoz-Avila ; Michael T. Cox
【Abstract】: Goal Driven Autonomy (GDA) is an agent model for reasoning about goals while acting in a dynamic environment. Since anomalous events may cause an agent's current goal to become invalid, GDA agents monitor the environment for such anomalies. When domains are both partially observable and dynamic, agents must reason about sensing and planning actions. Previous GDA work evaluated agents in domains that were partially observable, but does not address sensing actions with associated costs. Furthermore, partial observability still enabled generation of a grounded plan to reach the goal. We study agents where observability is more limited: the agent cannot generate a grounded plan because it does not know which future actions will be available until it explores more of the environment. We present a formalism of the problem that includes sensing costs, a GDA algorithm using this formalism, an examination of four methods of expectations under this formalism, and an implementation of the algorithm and empirical study.
【Keywords】:
【Paper Link】 【Pages】:2500-2506
【Authors】: Zipei Fan ; Ayumi Arai ; Xuan Song ; Apichon Witayangkurn ; Hiroshi Kanasugi ; Ryosuke Shibasaki
【Abstract】: Most of human mobility big datasets available by now, for example call detail records or twitter data with geotag, are always sparse and heavily biased. As a result, using such kind of data to directly represent real-world human mobility is unreliable and problematic. However, difficult though it is, a completion of human mobility turns out to be a promising way to minimize the issues of sparsity and bias. In this paper, we model the completion problem as a recommender system and therefore solve this problem in a collaborative filtering (CF) framework. We propose a spatio-temporal CF that simultaneously infers the topic distribution over users, time-of-days, days as well as locations, and then use the topic distributions to estimate a posterior over locations and infer the optimal location sequence in a Hidden Markov Model considering the spatio-temporal continuity. We apply and evaluate our algorithm using a real-world Call Detail Records dataset from Bangladesh and gives an application on Dynamic Census, which incorporates the survey data from cell phone users to generate an hourly population distribution with attributes.
【Keywords】:
【Paper Link】 【Pages】:2507-2513
【Authors】: Qingyu Guo ; Bo An ; Yair Zick ; Chunyan Miao
【Abstract】: Large scale smuggling of illegal goods is a long-standing problem, with $1.4b and thousands of agents assigned to protect the borders from such activity in the US-Mexico border alone. Illegal smuggling activities are usually blocked via inspection stations or ad-hoc checkpoints/roadblocks. Security resources are insufficient to man all stations at all times; furthermore, smugglers regularly conduct surveillance activities.This paper makes several contributions toward the challenging task of optimally interdicting an illegal network flow: i) A new Stackelberg game model for network flow interdiction; ii) A novel Column and Constraint Generation approach for computing the optimal defender strategy; iii) Complexity analysis of the column generation subproblem; iv) Compact convex nonlinear programs for solving the subproblems; v) Novel greedy and heuristic approaches for subproblems with good approximation guarantee. Experimental evaluation shows that our approach can obtain a robust enough solution outperforming the existing methods and heuristic baselines significantly and scale up to realistic-sized problems.
【Keywords】:
【Paper Link】 【Pages】:2514-2520
【Authors】: Ahmed Khalifa ; Aaron Isaksen ; Julian Togelius ; Andy Nealen
【Abstract】: We address the problem of making general video game playing agents play in a human-like manner. To this end, we introduce several modifications of the UCT formula used in Monte Carlo Tree Search that biases action selection towards repeating the current action, making pauses, and limiting rapid switching between actions. Playtraces of human players are used to model their propensity for repeated actions; this model is used for biasing the UCT formula. Experiments show that our modified MCTS agent, called BoT, plays quantitatively similar to human players as measured by the distribution of repeated actions. A survey of human observers reveals that the agent exhibits human-like playing style in some games but not others.
【Keywords】:
【Paper Link】 【Pages】:2521-2528
【Authors】: Zhaobin Kuang ; James A. Thomson ; Michael Caldwell ; Peggy L. Peissig ; Ron M. Stewart ; David Page
【Abstract】: Computational Drug Repositioning (CDR) is the knowledge discovery process of finding new indications for existing drugs leveraging heterogeneous drug-related data. Longitudinal observational data such as Electronic Health Records (EHRs) have become an emerging data source for CDR. To address the high-dimensional, irregular, subject and time-heterogeneous nature of EHRs, we propose Baseline Regularization (BR) and a variant that extend the one-way fixed effect model, which is a standard approach to analyze small-scale longitudinal data. For evaluation, we use the proposed methods to search for drugs that can lower Fasting Blood Glucose (FBG) level in the Marshfield Clinic EHR. Experimental results suggest that the proposed methods are capable of rediscovering drugs that can lower FBG level as well as identifying some potential blood sugar lowering drugs in the literature.
【Keywords】:
【Paper Link】 【Pages】:2529-2535
【Authors】: Sébastien Lallé ; Cristina Conati ; Giuseppe Carenini
【Abstract】: Confusion has been found to hinder user experience with visualizations. If confusion could be predicted and resolved in real time, user experience and satisfaction would greatly improve. In this paper, we focus on predicting occurrences of confusion during the interaction with a visualization using eye tracking and mouse data. The data was collected during a user study with ValueChart, an interactive visualization to support preferential choices. We report very promising results based on Random Forest classifiers.
【Keywords】:
【Paper Link】 【Pages】:2536-2543
【Authors】: Tuan M. V. Le ; Hady Wirawan Lauw
【Abstract】: Word cloud is a visualization form for text that is recognized for its aesthetic, social, and analytical values. Here, we are concerned with deepening its analytical value for visual comparison of documents. To aid comparative analysis of two or more documents, users need to be able to perceive similarities and differences among documents through their word clouds. However, as we are dealing with text, approaches that treat words independently may impede accurate discernment of similarities among word clouds containing different words of related meanings. We therefore motivate the principle of displaying related words in a coherent manner, and propose to realize it through modeling the latent aspects of words. Our WORD FLOCK solution brings together latent variable analysis for embedding and aspect modeling, and calibrated layout algorithm within a synchronized word cloud generation framework. We present the quantitative and qualitative results on real-life text corpora, showcasing how the word clouds are useful in preserving the information content of documents so as to allow more accurate visual comparison of documents.
【Keywords】:
【Paper Link】 【Pages】:2544-2552
【Authors】: Wonsung Lee ; Youngmin Lee ; Heeyoung Kim ; Il-Chul Moon
【Abstract】: Phenotyping with electronic health records (EHR) has received much attention in recent years because the phenotyping opens a new way to discover clinically meaningful insights, such as disease progression and disease subtypes without human supervisions. In spite of its potential benefits, the complex nature of EHR often requires more sophisticated methodologies compared with traditional methods. Previous works on EHR-based phenotyping utilized unsupervised and supervised learning methods separately by independently detecting phenotypes and predicting medical risk scores. To improve EHR-based phenotyping by bridging the separated methods, we present Bayesian nonparametric collaborative topic Poisson factorization (BN-CTPF) that is the first nonparametric content-based Poisson factorization and first application of jointly analyzing the phenotye topics and estimating the individual risk scores. BN-CTPF shows better performances in predicting the risk scores when we compared the model with previous matrix factorization and topic modeling methods including a Poisson factorization and its collaborative extensions. Also, BN-CTPF provides faceted views on the phenotype topics by patients' demographics. Finally, we demonstrate a scalable stochastic variational inference algorithm by applying BN-CTPF to a national-scale EHR dataset.
【Keywords】:
【Paper Link】 【Pages】:2553-2559
【Authors】: Liangda Li ; Hongyuan Zha
【Abstract】: In energy conservation research, energy disaggregation becomes an increasingly critical task, which takes a whole home electricity signal and decomposes it into its component appliances. While householder's daily energy usage behavior acts as one powerful cue for breaking down the entire household's energy consumption, existing works rarely modeled it straightforwardly. Instead, they either ignored the influence between users' energy usage behaviors, or modeled the influence between the energy usages of appliances.With ambiguous appliance usage membership of householders, we find it difficult to appropriately model the influence between appliances, since such influence is determined by human behaviors in energy usage. To address this problem, we propose to model the influence between householders' energy usage behaviors directly through a novel probabilistic model, which combines topic models with the Hawkes processes. The proposed model simultaneously disaggregates the whole home electricity signal into each component appliance and infers the appliance usage membership of household members, and enables those two tasks mutually benefit each other. Experimental results on both synthetic data and four real world data sets demonstrate the effectiveness of our model, which outperforms state-of-the-art approaches in not only decomposing the entire consumed energy to each appliance in houses, but also the inference of household structures. We further analyze the inferred appliance-householder assignment and the corresponding influence within the appliance usage of each householder and across different householders, which provides insight into appealing human behavior patterns in appliance usage.
【Keywords】:
【Paper Link】 【Pages】:2560-2567
【Authors】: Zhen Li ; Yizhou Yu
【Abstract】: Protein secondary structure prediction is an important problem in bioinformatics. Inspired by the recent successes of deep neural networks, in this paper, we propose an end-to-end deep network that predicts protein secondary structures from integrated local and global contextual features. Our deep architecture leverages convolutional neural networks with different kernel sizes to extract multiscale local contextual features. In addition, considering long-range dependencies existing in amino acid sequences, we set up a bidirectional neural network consisting of gated recurrent unit to capture global contextual features. Furthermore, multi-task learning is utilized to predict secondary structure labels and amino-acid solvent accessibility simultaneously. Our proposed deep network demonstrates its effectiveness by achieving state-of-the-art performance, i.e., 69.7% Q8 accuracy on the public benchmark CB513, 76.9% Q8 accuracy on CASP10 and 73.1% Q8 accuracy on CASP11. Our model and results are publicly available.
【Keywords】:
【Paper Link】 【Pages】:2568-2575
【Authors】: Si Liu ; Xinyu Ou ; Ruihe Qian ; Wei Wang ; Xiaochun Cao
【Abstract】: In this paper, we propose a novel Deep Localized Makeup Transfer Network to automatically recommend the most suitable makeup for a female and synthesis the makeup on her face. Given a before-makeup face, her most suitable makeup is determined automatically. Then, both the before makeup and the reference faces are fed into the proposed Deep Transfer Network to generate the after-makeup face. Our end-to-end makeup transfer network have several nice properties including: (1) with complete functions: including foundation, lip gloss, and eye shadow transfer; (2) cosmetic specific: different cosmetics are transferred in different manners; (3) localized: different cosmetics are applied on different facial regions; (4) producing naturally looking results without obvious artifacts; (5) controllable makeup lightness: various results from light makeup to heavy makeup can be generated. Qualitative and quantitative experiments show that our network performs much better than the methods of [Guo and Sim, 2009] and two variants of NerualStyle [Gatys et al., 2015a].
【Keywords】:
【Paper Link】 【Pages】:2576-2581
【Authors】: Ye Liu ; Yu Zheng ; Yuxuan Liang ; Shuming Liu ; David S. Rosenblum
【Abstract】: Urban water quality is of great importance to our daily lives. Prediction of urban water quality help control water pollution and protect human health. In this work, we forecast the water quality of a station over the next few hours, using a multi-task multi-view learning method to fuse multiple datasets from different domains. In particular, our learning model comprises two alignments. The first alignment is the spaio-temporal view alignment, which combines local spatial and temporal information of each station. The second alignment is the prediction alignment among stations, which captures their spatial correlations and performs co-predictions by incorporating these correlations. Extensive experiments on real-world datasets demonstrate the effectiveness of our approach.
【Keywords】:
【Paper Link】 【Pages】:2583-2589
【Authors】: Gautier Marti ; Sébastien Andler ; Frank Nielsen ; Philippe Donnat
【Abstract】: Researchers have used from 30 days to several years of daily returns as source data for clustering financial time series based on their correlations. This paper sets up a statistical framework to study the validity of such practices. We first show that clustering correlated random variables from their observed values is statistically consistent. Then, we also give a first empirical answer to the much debated question: How long should the time series be? If too short, the clusters found can be spurious; if too long, dynamics can be smoothed out.
【Keywords】:
【Paper Link】 【Pages】:2590-2596
【Authors】: Wookhee Min ; Bradford W. Mott ; Jonathan P. Rowe ; Barry Liu ; James C. Lester
【Abstract】: Recent years have seen a growing interest in player modeling for digital games. Goal recognition, which aims to accurately recognize players' goals from observations of low-level player actions, is a key problem in player modeling. However, player goal recognition poses significant challenges because of the inherent complexity and uncertainty pervading gameplay. In this paper, we formulate player goal recognition as a sequence labeling task and introduce a goal recognition framework based on long short-term memory (LSTM) networks. Results show that LSTM-based goal recognition is significantly more accurate than previous state-of-the-art methods, including n-gram encoded feedforward neural networks pre-trained with stacked denoising autoencoders, as well as Markov logic network-based models. Because of increased goal recognition accuracy and the elimination of labor-intensive feature engineering, LSTM-based goal recognition provides an effective solution to a central problem in player modeling for open-world digital games.
【Keywords】:
【Paper Link】 【Pages】:2597-2603
【Authors】: Anastasia Moskvina ; Jiamou Liu
【Abstract】: Creating new ties in a social network facilitates knowledge exchange and affects positional advantage. In this paper, we study the process, which we call network building, of establishing ties between two existing social networks in order to reach certain structural goals. We focus on the case when one of the two networks consists only of a single member and motivate this case from two perspectives. The first perspective is socialization: we ask how a newcomer can forge relationships with an existing network to place herself at the center. We prove that obtaining optimal solutions to this problem is NP-complete, and present several efficient algorithms to solve this problem and compare them with each other. The second perspective is network expansion: we investigate how a network may preserve or reduce its diameter through linking with a new node, hence ensuring small distance between its members. For both perspectives the experiment demonstrates that a small number of new links is usually sufficient to reach the respective goal.
【Keywords】:
【Paper Link】 【Pages】:2604-2610
【Authors】: Marco Ragni ; Christian Eichhorn ; Gabriele Kern-Isberner
【Abstract】: Human answer patterns in psychological reasoning experiments systematically deviate from predictions of classical logic. When interactions between any artificial reasoning system and humans are necessary this difference can be useful in some cases and lead to problems in other cases. Hence, other approaches than classical logic might be better suited to capture human inference processes. Evaluations are rare of how good such other approaches, e.g., non-monotonic logics, can explain psychological findings. In this article we consider the so-called Suppression Task, a core example in cognitive science about human reasoning that demonstrates that some additional information can lead to the suppression of simple inferences like the modus ponens. The psychological findings for this task have often been replicated and demonstrate a key-effect of human inferences. We analyze inferences of selected formal approaches and compare them by their capacity to cover human inference observed in the Suppression Task. A discussion on formal properties of successful theories conclude the paper.
【Keywords】:
【Paper Link】 【Pages】:2611-2617
【Authors】: Victor Shnayder ; Rafael M. Frongillo ; David C. Parkes
【Abstract】: Peer prediction is the problem of eliciting private, but correlated, information from agents. By rewarding an agent for the amount that their report "predicts" that of another agent, mechanisms can promote effort and truthful reports. A common concern in peer prediction is the multiplicity of equilibria, perhaps including high-payoff equilibria that reveal no information. Rather than assume agents counter-speculate and compute an equilibrium, we adopt replicator dynamics as a model for population learning. We take the size of the basin of attraction of the truthful equilibrium as a proxy for the robustness of truthful play. We study different mechanism designs, using models estimated from real peer evaluations in several massive on-line courses. Among other observations, we confirm that recent mechanisms present a significant improvement in robustness over earlier approaches.
【Keywords】:
【Paper Link】 【Pages】:2618-2624
【Authors】: Xuan Song ; Hiroshi Kanasugi ; Ryosuke Shibasaki
【Abstract】: Traffic congestion causes huge economic loss worldwide in every year due to wasted fuel, excessive air pollution, lost time, and reduced productivity. Understanding how humans move and select the transportation mode throughout a large-scale transportation network is vital for urban congestion prediction and transportation scheduling. In this study, we collect big and heterogeneous data (e.g., GPS records and transportation network data), and we build an intelligent system, namely DeepTransport, for simulating and predicting human mobility and transportation mode at a citywide level. The key component of DeepTransport is based on the deep learning architecture that that aims to understand human mobility and transportation patterns from big and heterogeneous data. Based on the learning model, given any time period, specific location of the city or people's observed movements, our system can automatically simulate or predict the persons' future movements and their transportation mode in the large-scale transportation network. Experimental results and validations demonstrate the efficiency and superior performance of our system, and suggest that human transportation mode may be predicted and simulated more easily than previously thought.
【Keywords】:
【Paper Link】 【Pages】:2625-2632
【Authors】: Yale Song ; Randall Davis ; Kaichen Ma ; Dana L. Penney
【Abstract】: We describe a sketch interpretation system that detects and classifies clock numerals created by subjects taking the Clock Drawing Test, a clinical tool widely used to screen for cognitive impairments (e.g., dementia). We describe how it balances appearance and context, and document its performance on some 2,000 drawings (about 24K clock numerals) produced by a wide spectrum of patients. We calibrate the utility of different forms of context, describing experiments with Conditional Random Fields trained and tested using a variety of features. We identify context that contributes to interpreting otherwise ambiguous or incomprehensible strokes. We describe ST-slices, a novel representation that enables "unpeeling" the layers of ink that result when people overwrite, which often produces ink impossible to analyze if only the final drawing is examined. We characterize when ST-slices work, calibrate their impact on performance, and consider their breadth of applicability.
【Keywords】:
【Paper Link】 【Pages】:2633-2639
【Authors】: Jakob Suchan ; Mehul Bhatt
【Abstract】: We present a computational framework for the grounding and semantic interpretation of dynamic visuo-spatial imagery consisting of video and eye-tracking data. Driven by cognitive film studies and visual perception research, we demonstrate key technological capabilities aimed at investigating attention and recipient effects vis-a-vis the motion picture; this encompasses high-level analysis of subject's visual fixation patterns and correlating this with (deep) semantic analysis of the dynamic visual data (e.g., fixation on movie characters, influence of cinematographic devices such as cuts). The framework and its application as a general AI-based assistive technology platform — integrating vision and KR — for cognitive film studies is highlighted.
【Keywords】:
【Paper Link】 【Pages】:2640-2646
【Authors】: Yunzhi Tan ; Min Zhang ; Yiqun Liu ; Shaoping Ma
【Abstract】: The performance of a recommendation system relies heavily on the feedback of users. Most of the traditional recommendation algorithms based only on historical ratings will encounter several difficulties given the problem of data sparsity. Users' feedback usually contains rich textual reviews in addition to numerical ratings. In this paper, we exploit textual review information, as well as ratings, to model user preferences and item features in a shared topic space and subsequently introduce them into a matrix factorization model for recommendation. To this end, the data sparsity problem is alleviated and good interpretability of the recommendation results is gained. Another contribution of this work is that we model the item feature distributions with rating-boosted reviews which combine textual reviews with user sentiments. Experimental results on 26 real-world datasets from Amazon demonstrate that our approach significantly improves the rating prediction accuracy compared with various state-of-the-art models, such as LFM, HFT, CTR and RMR models. And much higher improvement is achieved for users who have few ratings, which verifies the effectiveness of the proposed approach for sparse data. Moreover, our method also benefits much from reviews on top-N recommendation tasks.
【Keywords】:
【Paper Link】 【Pages】:2647-2653
【Authors】: Shu Tian ; Wei-Yi Pei ; Ze-Yu Zuo ; Xu-Cheng Yin
【Abstract】: There are a variety of grand challenges for text extraction in scene videos by robots and users, e.g., heterogeneous background, varied text, nonuniform illumination, arbitrary motion and poor contrast. Most previous video text detection methods are investigated with local information, i.e., within individual frames, with limited performance. In this paper, we propose a unified tracking based text detection system by learning locally and globally, which uniformly integrates detection, tracking, recognition and their interactions. In this system, scene text is first detected locally in individual frames. Second, an optimal tracking trajectory is learned and linked globally with all detection, recognition and prediction information by dynamic programming. With the tracking trajectory, final detection and tracking results are simultaneously and immediately obtained. Moreover, our proposed techniques are extensively evaluated on several public scene video text databases, and are much better than the state-of-the-art methods.
【Keywords】:
【Paper Link】 【Pages】:2654-2660
【Authors】: Kewei Tu
【Abstract】: Stochastic And-Or grammars (AOG) extend traditional stochastic grammars of language to model other types of data such as images and events. In this paper we propose a representation framework of stochastic AOGs that is agnostic to the type of the data being modeled and thus unifies various domain-specific AOGs. Many existing grammar formalisms and probabilistic models in natural language processing, computer vision, and machine learning can be seen as special cases of this framework. We also propose a domain-independent inference algorithm of stochastic context-free AOGs and show its tractability under a reasonable assumption. Furthermore, we provide two interpretations of stochastic context-free AOGs as a subset of probabilistic logic, which connects stochastic AOGs to the field of statistical relational learning and clarifies their relation with a few existing statistical relational models.
【Keywords】:
【Paper Link】 【Pages】:2661-2668
【Authors】: Ke Wang ; Zhendong Su
【Abstract】: Mathematical Word Problems (MWPs) are important for training students' literacy and numeracy skills. Traditionally MWPs have been manually designed; an effective automated MWP generator can significantly benefit education and research. The goal of this work is to efficiently synthesize MWPs that are authentic (i.e., similar to manually written problems), diverse (i.e., covering a wide range of mathematical tasks), and configurable (i.e., varying difficulty levels and solution characteristics). This is challenging because a generated problem needs to both exhibit a well-founded mathematical structure and also an easily understood natural language story. Our key insight is to leverage the important role that dimensional units play in MWPs, both textually and symbolically. We first synthesize a dimensionally consistent equation and then compose the natural language story via a bottom-up traversal of the equation tree. We have realized our technique and extensively evaluated its efficiency and effectiveness. Results show that the system can generate hundreds of valid problems per second with varying levels of difficulty. More importantly, we show, via a user study with 30 students from a local middle school, that the generated problems are statistically indistinguishable from actual textbook problems for practice and examination.
【Keywords】:
【Paper Link】 【Pages】:2669-2675
【Authors】: Zheng Wang ; Ruimin Hu ; Yi Yu ; Junjun Jiang ; Chao Liang ; Jinqiao Wang
【Abstract】: Person re-identification, as an important task in video surveillance and forensics applications, has been widely studied. But most of previous approaches are based on the key assumption that images for comparison have the same resolution and a uniform scale. Some recent works investigate how to match low resolution query images against high resolution gallery images, but still assume that the low-resolution query images have the same scale. In real scenarios, person images may not only be with low-resolution but also have different scales. Through investigating the distance variation behavior by changing image scales, we observe that scale-distance functions, generated by image pairs under different scales from the same person or different persons, are distinguishable and can be classified as feasible (for a pair of images from the same person) or infeasible (for a pair of images from different persons). The scale-distance functions are further represented by parameter vectors in the scale-distance function space. On this basis, we propose to learn a discriminating surface separating these feasible and infeasible functions in the scale-distance function space, and use it for reidentifying persons. Experimental results on two simulated datasets and one public dataset demonstrate the effectiveness of the proposed framework.
【Keywords】:
【Paper Link】 【Pages】:2676-2682
【Authors】: Shuai Xiao ; Junchi Yan ; Changsheng Li ; Bo Jin ; Xiangfeng Wang ; Xiaokang Yang ; Stephen M. Chu ; Hongyuan Zha
【Abstract】: Evaluating a scientist's past and future potential impact is key in decision making concerning with recruitment and funding, and is increasingly linked to publication citation count. Meanwhile, timely identifying those valuable work with great potential before they receive wide recognition and become highly cited Abstracts is both useful for readers and authors in many regards. We propose a method for predicting the citation counts of individual publications, over an arbitrary time period. Our approach explores paper-specific covariates, and a point process model to account for the aging effect and triggering role of recent citations, through which Abstracts lose and gain their popularity, respectively. Empirical results on the Microsoft Academic Graph data suggests that our model can be useful for both prediction and interpretability.
【Keywords】:
【Paper Link】 【Pages】:2683-2689
【Authors】: Xiao-Feng Xie ; Zun-Jing Wang
【Abstract】: In this paper, we present a method using AI techniques to solve a case of pure mathematics applications for finding narrow admissible tuples. The original problem is formulated into a combinatorial optimization problem. In particular, we show how to exploit the local search structure to formulate the problem landscape for dramatic reductions in search space and for non-trivial elimination in search barriers, and then to realize intelligent search strategies for effectively escaping from local minima. Experimental results demonstrate that the proposed method is able to efficiently find best known solutions. This research sheds light on exploiting the local problem structure for an efficient search in combinatorial landscapes as an application of AI to a new problem domain.
【Keywords】:
【Paper Link】 【Pages】:2690-2696
【Authors】: Junchi Yan ; Shuai Xiao ; Changsheng Li ; Bo Jin ; Xiangfeng Wang ; Bin Ke ; Xiaokang Yang ; Hongyuan Zha
【Abstract】: Merger and Acquisition (M&A) has been a critical practice about corporate restructuring. Previous studies are mostly devoted to evaluating the suitability of M&A between a pair of investor and target company, or a target company for its propensity of being acquired. This paper focuses on the dual problem of predicting an investor's prospective M&A based on its activities and firmographics. We propose to use a mutually-exciting point process with a regression prior to quantify the investor's M&A behavior. Our model is motivated by the so-called contagious 'wave-like' M&A phenomenon, which has been well-recognized by the economics and management communities. A tailored model learning algorithm is devised that incorporates both static profile covariates and past M&A activities. Results on CrunchBase suggest the superiority of our model. The collected dataset and code will be released together with the paper.
【Keywords】:
【Paper Link】 【Pages】:2697-2703
【Authors】: Dingqi Yang ; Bin Li ; Philippe Cudré-Mauroux
【Abstract】: Capturing place semantics is critical for enabling location-based applications. Techniques for assigning semantic labels (e.g., "bar" or "office") to unlabeled places mainly resort to mining user activity logs by exploiting visiting patterns. However, existing approaches focus on inferring place labels with a static user activity dataset, and ignore the visiting pattern dynamics in user activity streams, leading to the rapid decrease of labeling accuracy over time. In this paper, we tackle the problem of semantic place labeling over user activity streams. We formulate this problem as a classification problem by characterizing each place through its fine-grained visiting patterns, which encode the visiting frequency of each user in each typical time slot. However, with the incoming activities of new users in data streams, such fine-grained visiting patterns constantly grow, leading to a continuously expanding feature space. To solve this issue, we propose an updatable sketching technique that creates and incrementally updates a set of compact and fixed-size sketches to approximate the similarity between fine-grained visiting patterns of ever-growing size. We further consider the discriminative weights of user activities in place labeling, and seamlessly incorporate them into our sketching method. Our empirical evaluation on real-world datasets demonstrates the validity of our approach and shows that sketches can be efficiently and effectively used to infer place labels over user activity streams.
【Keywords】:
【Paper Link】 【Pages】:2704-2710
【Authors】: Xiuwen Yi ; Yu Zheng ; Junbo Zhang ; Tianrui Li
【Abstract】: Many sensors have been deployed in the physical world, generating massive geo-tagged time series data. In reality, we usually lose readings of sensors at some unexpected moments because of sensor or communication errors. Those missing readings do not only affect real-time monitoring but also compromise the performance of further data analysis. In this paper, we propose a spatio-temporal multi-view-based learning (ST-MVL) method to collectively fill missing readings in a collection of geo-sensory time series data, considering 1) the temporal correlation between readings at different timestamps in the same series and 2) the spatial correlation between different time series. Our method combines empirical statistic models, consisting of Inverse Distance Weighting and Simple Exponential Smoothing, with data-driven algorithms, comprised of User-based and Item-based Collaborative Filtering. The former models handle the general missing cases based on empirical assumptions derived from history data over a long period, standing for two global views from a spatial and temporal perspective respectively. The latter algorithms deal with special cases where empirical assumptions may not hold, based on recent contexts of data, denoting two local views from a spatial and temporal perspective respectively. The predictions of the four views are aggregated to a final value in a multi-view learning algorithm. We evaluate our method based on Beijing air quality and meteorological data, finding our model's advantages beyond ten baseline approaches.
【Keywords】:
【Paper Link】 【Pages】:2711-2717
【Authors】: Xiang Yu ; Zhe L. Lin ; Shaoting Zhang ; Dimitris N. Metaxas
【Abstract】: Non-linear regression is a fundamental and yet under-developing methodology in solving many problems in Artificial Intelligence. The canonical control and predictions mostly utilize linear models or multi-linear models. However, due to the high non-linearity of the systems, those linear prediction models cannot fully cover the complexity of the problems. In this paper, we propose a robust two-stage hierarchical regression approach, to solve a popular Human-Computer Interaction, the unconstrained face-in-the-wild keypoint detection problem for computers. The environment is the still images, videos and live camera streams from machine vision. We firstly propose a holistic regression model to initialize the face fiducial points under different head pose assumptions. Second, to reduce local shape variance, a hierarchical part-based regression method is further proposed to refine the global regression output. Experiments on several challenging faces-in-the-wild datasets demonstrate the consistently better accuracy of our method, when compared to the state-of-the-art.
【Keywords】:
【Paper Link】 【Pages】:2718-2724
【Authors】: Lu Zhang ; Yongkai Wu ; Xintao Wu
【Abstract】: Discrimination discovery is to unveil discrimination against a specific individual by analyzing the historical dataset. In this paper, we develop a general technique to capture discrimination based on the legally grounded situation testing methodology. For any individual, we find pairs of tuples from the dataset with similar characteristics apart from belonging or not to the protected-by-law group and assign them in two groups. The individual is considered as discriminated if significant difference is observed between the decisions from the two groups. To find similar tuples, we make use of the Causal Bayesian Networks and the associated causal inference as a guideline. The causal structure of the dataset and the causal effect of each attribute on the decision are used to facilitate the similarity measurement. Through empirical assessments on a real dataset, our approach shows good efficacy both in accuracy and efficiency.
【Keywords】:
【Paper Link】 【Pages】:2725-2731
【Authors】: Ruohan Zhang ; Zhao Song
【Abstract】: We introduce the Maximum Sustainable Yield problem for a multi-robot foraging and construction system, inspired by the relationship between the natural resource growth and harvesting behaviors in an ecosystem. The resources spawn according to the logistic model and are vulnerable to overharvesting. The robots must maintain sustainability while maximizing productivity. The foraging robots harvest different types of resources, which enable a construction robot to build new foraging robots. We design algorithms to perform robot construction, assignment, and scheduling. We propose an adaptive algorithm to overcome the problem that resource growth model is often unknown. We demonstrate that our algorithms are robust to noises in the actuation and the environment. The case where the observation noise could harm sustainability is discussed.
【Keywords】:
【Paper Link】 【Pages】:2732-2739
【Authors】: Wei-Long Zheng ; Bao-Liang Lu
【Abstract】: Individual differences across subjects and non-stationary characteristic of electroencephalography (EEG) limit the generalization of affective brain-computer interfaces in real-world applications. On the other hand, it is very time consuming and expensive to acquire a large number of subject-specific labeled data for learning subject-specific models. In this paper, we propose to build personalized EEG-based affective models without labeled target data using transfer learning techniques. We mainly explore two types of subject-to-subject transfer approaches. One is to exploit shared structure underlying source domain (source subject) and target domain (target subject). The other is to train multiple individual classifiers on source subjects and transfer knowledge about classifier parameters to target subjects, and its aim is to learn a regression function that maps the relationship between feature distribution and classifier parameters. We compare the performance of five different approaches on an EEG dataset for constructing an affective model with three affective states: positive, neutral, and negative. The experimental results demonstrate that our proposed subject transfer framework achieves the mean accuracy of 76.31% in comparison with a conventional generic classifier with 56.73% in average.
【Keywords】:
【Paper Link】 【Pages】:2740-2746
【Authors】: Siddhartha Banerjee ; Prasenjit Mitra
【Abstract】: The growth of Wikipedia, limited by the availability of knowledgeable authors, cannot keep pace with the ever increasing requirements and demands of the readers. In this work, we propose WikiWrite, a system capable of generating content for new Wikipedia articles automatically. First, our technique obtains feature representations of entities on Wikipedia. We adapt an existing work on document embeddings to obtain vector representations of words and paragraphs. Using the representations, we identify articles that are very similar to the new entity on Wikipedia. We train machine learning classifiers using content from the similar articles to assign web retrieved content on the new entity into relevant sections in the Wikipedia article. Second, we propose a novel abstractive summarization technique that uses a two-step integer-linear programming (ILP) model to synthesize the assigned content in each section and rewrite the content to produce a well-formed informative summary. Our experiments show that our technique is able to reconstruct existing articles in Wikipedia with high accuracies. We also create several articles using our approach in the English Wikipedia, most of which have been retained in the online encyclopedia.
【Keywords】:
【Paper Link】 【Pages】:2747-2753
【Authors】: Emanuele Bastianelli ; Danilo Croce ; Andrea Vanzo ; Roberto Basili ; Daniele Nardi
【Abstract】: Spoken Language Understanding in Interactive Robotics provides computational models of human-machine communication based on the vocal input. However, robots operate in specific environments and the correct interpretation of the spoken sentences depends on the physical, cognitive and linguistic aspects triggered by the operational environment. Grounded language pro- cessing should exploit both the physical constraints of the context as well as knowledge assumptions of the robot. These include the subjective perception of the environment that explicitly affects linguistic reasoning. In this work, a standard linguistic pipeline for semantic parsing is extended toward a form of perceptually informed natural language processing that combines discriminative learning and distributional semantics. Empirical results achieve up to a 40% of relative error reduction.
【Keywords】:
【Paper Link】 【Pages】:2754-2760
【Authors】: Qian Chen ; Xiao-Dan Zhu ; Zhen-Hua Ling ; Si Wei ; Hui Jiang
【Abstract】: Distributed representation learned with neural networks has recently shown to be effective in modeling natural languages at fine granularities such as words, phrases, and even sentences. Whether and how such an approach can be extended to help model larger spans of text, e.g., documents, is intriguing, and further investigation would still be desirable. This paper aims to enhance neural network models for such a purpose. A typical problem of document-level modeling is automatic summarization, which aims to model documents in order to generate summaries. In this paper, we propose neural models to train computers not just to pay attention to specific regions and content of input documents with attention models, but also distract them to traverse between different content of a document so as to better grasp the overall meaning for summarization. Without engineering any features, we train the models on two large datasets. The models achieve the state-of-the-art performance, and they significantly benefit from the distraction modeling, particularly when input documents are long.
【Keywords】:
【Paper Link】 【Pages】:2761-2767
【Authors】: Yong Cheng ; Shiqi Shen ; Zhongjun He ; Wei He ; Hua Wu ; Maosong Sun ; Yang Liu
【Abstract】: The attentional mechanism has proven to be effective in improving end-to-end neural machine translation. However, due to the intricate structural divergence between natural languages, unidirectional attention-based models might only capture partial aspects of attentional regularities. We propose agreement-based joint training for bidirectional attention-based end-to-end neural machine translation. Instead of training source-to-target and target-to-source translation models independently, our approach encourages the two complementary models to agree on word alignment matrices on the same training data. Experiments on Chinese-English and English-French translation tasks show that agreement-based joint training significantly improves both alignment and translation quality over independent training.
【Keywords】:
【Paper Link】 【Pages】:2768-2774
【Authors】: Shamil Chollampatt ; Kaveh Taghipour ; Hwee Tou Ng
【Abstract】: Phrase-based statistical machine translation (SMT) systems have previously been used for the task of grammatical error correction (GEC) to achieve state-of-the-art accuracy. The superiority of SMT systems comes from their ability to learn text transformations from erroneous to corrected text, without explicitly modeling error types. However, phrase-based SMT systems suffer from limitations of discrete word representation, linear mapping, and lack of global context. In this paper, we address these limitations by using two different yet complementary neural network models, namely a neural network global lexicon model and a neural network joint model. These neural networks can generalize better by using continuous space representation of words and learn non-linear mappings. Moreover, they can leverage contextual information from the source sentence more effectively. By adding these two components, we achieve statistically significant improvement in accuracy for grammatical error correction over a state-of-the-art GEC system.
【Keywords】:
【Paper Link】 【Pages】:2775-2781
【Authors】: Lingjia Deng ; Janyce Wiebe
【Abstract】: Recognizing sources of opinions is an important task in sentiment analysis. Different from previous works which categorize an opinion according to whether the source is the writer or the source is a noun phrase, we propose a new categorization of opinions according to the role that the source plays. The source of a participant opinion is a participant in the event that triggers the opinion. On the contrary, the source of a non-participant opinion is not a participant. Based on this new categorization, we classify an opinion using phrase-level embeddings. A transductive learning method is used for the classifier since there is no existing annotated corpora of this new categorization. A joint prediction model of Probabilistic Soft Logic then recognizes the sources of the two types of opinions in a single model. The experiments have shown that our model improves recognizing sources of opinions over baselines and several state-of-the-art works.
【Keywords】:
【Paper Link】 【Pages】:2782-2788
【Authors】: Yuyun Gong ; Qi Zhang
【Abstract】: Along with the increasing requirements, the hashtag recommendation task for microblogs has been receiving considerable attention in recent years. Various researchers have studied the problem from different aspects. However, most of these methods usually need handcrafted features. Motivated by the successful use of convolutional neural networks (CNNs) for many natural language processing tasks, in this paper, we adopt CNNs to perform the hashtag recommendation problem. To incorporate the trigger words whose effectiveness have been experimentally evaluated in several previous works, we propose a novel architecture with an attention mechanism. The results of experiments on the data collected from a real world microblogging service demonstrated that the proposed model outperforms state-of-the-art methods. By incorporating trigger words into the consideration, the relative improvement of the proposed method over the state-of-the-art method is around 9.4% in the F1-score.
【Keywords】:
【Paper Link】 【Pages】:2789-2795
【Authors】: Lin Gui ; Ruifeng Xu ; Yulan He ; Qin Lu ; Zhongyu Wei
【Abstract】: Intersubjectivity is an important concept in psychology and sociology. It refers to sharing conceptualizations through social interactions in a community and using such shared conceptualization as a resource to interpret things that happen in everyday life. In this work, we make use of intersubjectivity as the basis to model shared stance and subjectivity for sentiment analysis. We construct an intersubjectivity network which links review writers, terms they used, as well as the polarities of the terms. Based on this network model, we propose a method to learn writer embeddings which are subsequently incorporated into a convolutional neural network for sentiment analysis. Evaluations on the IMDB, Yelp 2013 and Yelp 2014 datasets show that the proposed approach has achieved the state-of-the-art performance.
【Keywords】:
【Paper Link】 【Pages】:2796-2802
【Authors】: Homa B. Hashemi ; Rebecca Hwa
【Abstract】: Ungrammatical sentences present challenges for statistical parsers because the well-formed trees they produce may not be appropriate for these sentences. We introduce a framework for reviewing the parses of ungrammatical sentences and extracting the coherent parts whose syntactic analyses make sense. We call this task parse tree fragmentation. In this paper, we propose a training methodology for fragmenting parse trees without using a task-specific annotated corpus. We also propose some fragmentation strategies and compare their performance on an extrinsic task - fluency judgments in two domains: English-as-a-Second Language (ESL) and machine translation (MT). Experimental results show that the proposed fragmentation strategies are competitive with existing methods for making fluency judgments; they also suggest that the overall framework is a promising way to handle syntactically unusual sentences.
【Keywords】:
【Paper Link】 【Pages】:2803-2809
【Authors】: Duc Tam Hoang ; Shamil Chollampatt ; Hwee Tou Ng
【Abstract】: Grammatical error correction (GEC) is the task of detecting and correcting grammatical errors in texts written by second language learners. The statistical machine translation (SMT) approach to GEC, in which sentences written by second language learners are translated to grammatically correct sentences, has achieved state-of-the-art accuracy. However, the SMT approach is unable to utilize global context. In this paper, we propose a novel approach to improve the accuracy of GEC, by exploiting the n-best hypotheses generated by an SMT approach. Specifically, we build a classifier to score the edits in the n-best hypotheses. The classifier can be used to select appropriate edits or re-rank the n-best hypotheses. We apply these methods to a state-of-the-art GEC system that uses the SMT approach. Our experiments show that our methods achieve statistically significant improvements in accuracy over the best published results on a benchmark test dataset on GEC.
【Keywords】:
【Paper Link】 【Pages】:2810-2816
【Authors】: Jizhou Huang ; Shiqi Zhao ; Shiqiang Ding ; Haiyang Wu ; Mingming Sun ; Haifeng Wang
【Abstract】: Entity recommendation, providing entity suggestions relevant to the query that a user is searching for, has become a key feature of today's web search engine. Despite the fact that related entities are relevant to users' search queries, sometimes users cannot easily understand the recommended entities without evidences. This paper proposes a statistical model consisting of four sub-models to generate evidences for entities, which can help users better understand each recommended entity, and figure out the connections between the recommended entities and a given query. The experiments show that our method is domain independent, and can generate catchy and interesting evidences in the application of entity recommendation.
【Keywords】:
【Paper Link】 【Pages】:2817-2823
【Authors】: Shujian Huang ; Huifeng Sun ; Chengqi Zhao ; Jinsong Su ; Xin-Yu Dai ; Jiajun Chen
【Abstract】: Hierarchical phrase-based translation systems (HPBs) perform translation using a synchronous context free grammar which has only one unified non-terminal for every translation rule. While the usage of the unified non-terminal brings freedom to generate translations with almost arbitrary structures, it also takes the risks of generating low-quality translations which has a wrong syntactic structure. In this paper, we propose tree-state models to discriminate the good or bad usage of translation rules based on the syntactic structures of the source sentence. We propose to use statistical models and context dependent features to estimate the probability of each tree state for each translation rule and punish the usage of rules in the translation system which violates their tree states. Experimental results demonstrate that these simple models could bring significant improvements to the translation quality.
【Keywords】:
【Paper Link】 【Pages】:2824-2830
【Authors】: Peng Jin ; Yue Zhang ; Xingyuan Chen ; Yunqing Xia
【Abstract】: Words are central to text classification. It has been shown that simple Naive Bayes models with word and bigram features can give highly competitive accuracies when compared to more sophisticated models with part-of-speech, syntax and semantic features. Embeddings offer distributional features about words. We study a conceptually simple classification model by exploiting multi-prototype word embeddings based on text classes. The key assumption is that words exhibit different distributional characteristics under different text classes. Based on this assumption, we train multi-prototype distributional word representations for different text classes. Given a new document, its text class is predicted by maximizing the probabilities of embedding vectors of its words under the class. In two standard classification benchmark datasets, one is balance and the other is imbalance, our model outperforms state-of-the-art systems, on both accuracy and macro-average F-1 score.
【Keywords】:
【Paper Link】 【Pages】:2831-2837
【Authors】: Hatim Khouzaimi ; Romain Laroche ; Fabrice Lefèvre
【Abstract】: In this article, reinforcement learning is used to learn an optimal turn-taking strategy for vocal human-machine dialogue. The Orange Labs' Majordomo dialogue system, which allows the users to have conversations within a smart home, has been upgraded to an incremental version. First, a user simulator is built in order to generate a dialogue corpus which thereafter is used to optimise the turn-taking strategy from delayed rewards with the Fitted-Q reinforcement learning algorithm. Real users test and evaluate the new learnt strategy, versus a non-incremental and a handcrafted incremental strategies. The data-driven strategy is shown to significantly improve the task completion ratio and to be preferred by the users according to subjective metrics.
【Keywords】:
【Paper Link】 【Pages】:2838-2844
【Authors】: Fei Li ; Yue Zhang ; Meishan Zhang ; Donghong Ji
【Abstract】: Extracting adverse drug events receives much research attention in the biomedical community. Previous work adopts pipeline models, firstly recognizing drug/disease entity mentions and then identifying adverse drug events from drug/disease pairs. In this paper, we investigate joint models for simultaneously extracting drugs, diseases and adverse drug events. Compared with pipeline models, joint models have two main advantages. First, they make use of information integration to facilitate performance improvement; second, they reduce error propagation in pipeline methods. We compare a discrete model and a deep neural model for extracting drugs, diseases and adverse drug events jointly. Experimental results on a standard ADE corpus show that the discrete joint model outperforms a state-of-the-art baseline pipeline significantly. In addition, when discrete features are replaced by neural features, the recall is further improved.
【Keywords】:
【Paper Link】 【Pages】:2845-2851
【Authors】: Xiang Li ; Lili Mou ; Rui Yan ; Ming Zhang
【Abstract】: Existing open-domain human-computer conversation systems are typically passive: they either synthesize or retrieve a reply provided with a human-issued utterance. It is generally presumed that humans should take the role to lead the conversation and introduce new content when a stalemate occurs, and that computers only need to "respond." In this paper, we propose STALEMATEBREAKER, a conversation system that can proactively introduce new content when appropriate. We design a pipeline to determine when, what, and how to introduce new content during human-computer conversation. We further propose a novel reranking algorithm Bi-PageRank-HITS to enable rich interaction between conversation context and candidate replies. Experiments show that both the content-introducing approach and the reranking algorithm are effective. Our full STALEMATEBREAKER model outperforms a state-of-the-practice conversation system by +14.4% p@1 when a stalemate occurs.
【Keywords】:
【Paper Link】 【Pages】:2852-2858
【Authors】: Xiaoqing Li ; Jiajun Zhang ; Chengqing Zong
【Abstract】: Neural Machine translation has shown promising results in recent years. In order to control the computational complexity, NMT has to employ a small vocabulary, and massive rare words outside the vocabulary are all replaced with a single unk symbol. Besides the inability to translate rare words, this kind of simple approach leads to much increased ambiguity of the sentences since meaningless unks break the structure of sentences, and thus hurts the translation and reordering of the in-vocabulary words. To tackle this problem, we propose a novel substitution-translation-restoration method. In substitution step, the rare words in a testing sentence are replaced with similar in-vocabulary words based on a similarity model learnt from monolingual data. In translation and restoration steps, the sentence will be translated with a model trained on new bilingual data with rare words replaced, and finally the translations of the replaced words will be substituted by that of original ones. Experiments on Chinese-to-English translation demonstrate that our proposed method can achieve more than 4 BLEU points over the attention-based NMT. When compared to the recently proposed method handling rare words in NMT, our method can also obtain an improvement by nearly 3 BLEU points.
【Keywords】:
【Paper Link】 【Pages】:2859-2865
【Authors】: Chen Liang ; Praveen K. Paritosh ; Vinodh Rajendran ; Kenneth D. Forbus
【Abstract】: Semantic similarity of text plays an important role in many NLP tasks. It requires using both local information like lexical semantics and structural information like syntactic structures. Recent progress in word representation provides good resources for lexical semantics, and advances in natural language analysis tools make it possible to efficiently generate syntactic and semantic annotations. However, how to combine them to capture the semantics of text is still an open question. Here, we propose a new alignment-based approach to learn semantic similarity. It uses a hybrid representation, attributed relational graphs, to encode lexical, syntactic and semantic information. Alignment of two such graphs combines local and structural information to support similarity estimation. To improve alignment, we introduced structural constraints inspired by a cognitive theory of similarity and analogy. Usually only similarity labels are given in training data and the true alignments are unknown, so we address the learning problem using two approaches: alignment as feature extraction and alignment as latent variable. Our approach is evaluated on the paraphrase identification task and achieved results competitive with the state-of-the-art.
【Keywords】:
【Paper Link】 【Pages】:2866-2872
【Authors】: Yankai Lin ; Zhiyuan Liu ; Maosong Sun
【Abstract】: Distributed knowledge representation (KR) encodes both entities and relations in a low-dimensional semantic space, which has significantly promoted the performance of relation extraction and knowledge reasoning. In many knowledge graphs (KG), some relations indicate attributes of entities (attributes) and others indicate relations between entities (relations). Existing KR models regard all relations equally, and usually suffer from poor accuracies when modeling one-to-many and many-to-one relations, mostly composed of attribute. In this paper, we distinguish existing KG-relations into attributes and relations, and propose a new KR model with entities, attributes and relations (KR-EAR). The experiment results show that, by special modeling of attribute, KR-EAR can significantly outperform state-of-the-art KR models in prediction of entities, attributes and relations.
【Keywords】:
【Paper Link】 【Pages】:2873-2879
【Authors】: Pengfei Liu ; Xipeng Qiu ; Xuanjing Huang
【Abstract】: Neural network based methods have obtained great progress on a variety of natural language processing tasks. However, in most previous works, the models are learned based on single-task supervised objectives, which often suffer from insufficient training data. In this paper, we use the multi-task learning framework to jointly learn across multiple related tasks. Based on recurrent neural network, we propose three different mechanisms of sharing information to model text with task-specific and shared layers. The entire network is trained jointly on all these tasks. Experiments on four benchmark text classification tasks show that our proposed models can improve the performance of a task with the help of other related tasks.
【Keywords】:
【Paper Link】 【Pages】:2880-2886
【Authors】: Yijia Liu ; Wanxiang Che ; Jiang Guo ; Bing Qin ; Ting Liu
【Abstract】: Many natural language processing (NLP) tasks can be generalized into segmentation problem. In this paper, we combine semi-CRF with neural network to solve NLP segmentation tasks. Our model represents a segment both by composing the input units and embedding the entire segment. We thoroughly study different composition functions and different segment embeddings. We conduct extensive experiments on two typical segmentation tasks: named entity recognition (NER) and Chinese word segmentation (CWS). Experimental results show that our neural semi-CRF model benefits from representing the entire segment and achieves the state-of-the-art performance on CWS benchmark dataset and competitive results on the CoNLL03 NER dataset.
【Keywords】:
【Paper Link】 【Pages】:2887-2893
【Authors】: Yijia Liu ; Wanxiang Che ; Bing Qin ; Ting Liu
【Abstract】: Standard incremental parsing algorithm employs a single scoring function and beam-search to find the best parse tree from an exponentially large search space. Inspired by recently proposed HC-search framework, we decompose the incremental parsing algorithm into two steps: first searching a set of high-quality outputs with beam-search, and second selecting the best output with a ranking model. We learn our incremental parsing model with a relaxed learning objective. We incorporate arbitrary features in our ranking model and learn the model from fine grain ranking examples. Experimental results on standard English and Chinese datasets show our method significantly outperforms a strong baseline.
【Keywords】:
【Paper Link】 【Pages】:2894-2900
【Authors】: Alessandro Raganato ; Claudio Delli Bovi ; Roberto Navigli
【Abstract】: The hyperlink structure of Wikipedia constitutes a key resource for many Natural Language Processing tasks and applications, as it provides several million semantic annotations of entities in context. Yet only a small fraction of mentions across the entire Wikipedia corpus is linked. In this paper we present the automatic construction and evaluation of a Semantically Enriched Wikipedia in which the overall number of linked mentions has been more than tripled solely by exploiting the structure of Wikipedia itself and the wide-coverage sense inventory of BabelNet. As a result we obtain a sense-annotated corpus with more than 200 million annotations of over 4 million different concepts and named entities. We then show that our corpus leads to competitive results on multiple tasks, such as Entity Linking and Word Similarity.
【Keywords】:
【Paper Link】 【Pages】:2901-2907
【Authors】: Yangqiu Song ; Shyam Upadhyay ; Haoruo Peng ; Dan Roth
【Abstract】: Dataless text classification [Chang et al., 2008] is a classification paradigm which maps documents into a given label space without requiring any annotated training data. This paper explores a cross-lingual variant of this paradigm, where documents in multiple languages are classified into an English label space. We use CLESA (cross-lingual explicit semantic analysis) to embed both foreign language documents and an English label space into a shared semantic space, and select the best label(s) for a document using the similarity between the corresponding semantic representations. We illustrate our approach by experimenting with classifying documents in 88 different languages into the same English label space. In particular, we show that CLESA is better than using a monolingual ESA on the target foreign language and translating the English labels into that language. Moreover, the evaluation on two benchmarks, TED and RCV2, showed that cross-lingual dataless classification outperforms supervised learning methods when a large collection of annotated documents is not available.
【Keywords】:
【Paper Link】 【Pages】:2908-2914
【Authors】: Michael Spranger ; Jakob Suchan ; Mehul Bhatt
【Abstract】: We present a system for generating and understanding of dynamic and static spatial relations in robotic interaction setups. Robots describe an environment of moving blocks using English phrases that include spatial relations such as "across" and "in front of." We evaluate the system in robot-robot interactions and show that the system can robustly deal with visual perception errors, language omissions and ungrammatical utterances.
【Keywords】:
【Paper Link】 【Pages】:2915-2921
【Authors】: Fei Sun ; Jiafeng Guo ; Yanyan Lan ; Jun Xu ; Xueqi Cheng
【Abstract】: Recently, Word2Vec tool has attracted a lot of interest for its promising performances in a variety of natural language processing (NLP) tasks. However, a critical issue is that the dense word representations learned in Word2Vec are lacking of interpretability. It is natural to ask if one could improve their interpretability while keeping their performances. Inspired by the success of sparse models in enhancing interpretability, we propose to introduce sparse constraint into Word2Vec. Specifically, we take the Continuous Bag of Words (CBOW) model as an example in our study and add the l1 regularizer into its learning objective. One challenge of optimization lies in that stochastic gradient descent (SGD) cannot directly produce sparse solutions with l1 regularizer in online training. To solve this problem, we employ the Regularized Dual Averaging (RDA) method, an online optimization algorithm for regularized stochastic learning. In this way, the learning process is very efficient and our model can scale up to very large corpus to derive sparse word representations. The proposed model is evaluated on both expressive power and interpretability. The results show that, compared with the original CBOW model, the proposed model can obtain state-of-the-art results with better interpretability using less than 10% non-zero elements.
【Keywords】:
【Paper Link】 【Pages】:2922-2928
【Authors】: Shengxian Wan ; Yanyan Lan ; Jun Xu ; Jiafeng Guo ; Liang Pang ; Xueqi Cheng
【Abstract】: Semantic matching, which aims to determine the matching degree between two texts, is a fundamental problem for many NLP applications. Recently, deep learning approach has been applied to this problem and significant improvements have been achieved. In this paper, we propose to view the generation of the global interaction between two texts as a recursive process: i.e.~the interaction of two texts at each position is a composition of the interactions between their prefixes as well as the word level interaction at the current position. Based on this idea, we propose a novel deep architecture, namely Match-SRNN, to model the recursive matching structure. Firstly, a tensor is constructed to capture the word level interactions. Then a spatial RNN is applied to integrate the local interactions recursively, with importance determined by four types of gates. Finally, the matching score is calculated based on the global interaction. We show that, after degenerated to the exact matching scenario, Match-SRNN can approximate the dynamic programming process of longest common subsequence. Thus, there exists a clear interpretation for Match-SRNN. Our experiments on two semantic matching tasks showed the effectiveness of Match-SRNN, and its ability of visualizing the learned matching structure.
【Keywords】:
【Paper Link】 【Pages】:2929-2925
【Authors】: Bingning Wang ; Shangmin Guo ; Kang Liu ; Shizhu He ; Jun Zhao
【Abstract】: Recently proposed machine comprehension (MC) application is an effort to deal with natural language understanding problem. However, the small size of machine comprehension labeled data confines the application of deep neural networks architectures that have shown advantage in semantic inference tasks. Previous methods use a lot of NLP tools to extract linguistic features but only gain little improvement over simple baseline. In this paper, we build an attention-based recurrent neural network model, train it with the help of external knowledge which is semantically relevant to machine comprehension, and achieves a new state-of-art result.
【Keywords】:
【Paper Link】 【Pages】:2936-2942
【Authors】: Chang Wang ; Liangliang Cao ; James Fan
【Abstract】: In this paper, we present a novel approach for relation extraction using only term pairs as the input without textual features. We aim to build a single joint space for each relation which is then used to produce relation specific term embeddings. The proposed method fits particularly well for domains in which similar arguments are often associated with similar relations. It can also handle the situation when the labeled data is limited. The proposed method is evaluated both theoretically with a proof for the closed-form solution and experimentally with promising results on both DBpedia and medical relations.
【Keywords】:
【Paper Link】 【Pages】:2943-2949
【Authors】: Qixin Wang ; Tianyi Luo ; Dong Wang ; Chao Xing
【Abstract】: Learning and generating Chinese poems is a charming yet challenging task. Traditional approaches involve various language modeling and machine translation techniques, however, they perform not as well when generating poems with complex pattern constraints, for example Song iambics, a famous type of poems that involve variable-length sentences and strict rhythmic patterns. This paper applies the attention-based sequence-to-sequence model to generate Chinese Song iambics. Specifically, we encode the cue sentences by a bi-directional Long-Short Term Memory (LSTM) model and then predict the entire iambic with the information provided by the encoder, in the form of an attention-based LSTM that can regularize the generation process by the fine structure of the input cues. Several techniques are investigated to improve the model, including global context integration, hybrid style training, character vector initialization and adaptation. Both the automatic and subjective evaluation results show that our model indeed can learn the complex structural and rhythmic patterns of Song iambics, and the generation is rather successful.
【Keywords】:
【Paper Link】 【Pages】:2950-2956
【Authors】: Rui Wang ; Hai Zhao ; Sabine Ploux ; Bao-Liang Lu ; Masao Utiyama
【Abstract】: Most existing bilingual embedding methods for Statistical Machine Translation (SMT) suffer from two obvious drawbacks. First, they only focus on simple context such as word count and co-occurrence in document or sliding window to build word embedding, ignoring latent useful information from selected context. Second, word sense but not word form is supposed to be the minimal semantic unit while most existing works are still for word representation. This paper presents Bilingual Graph-based Semantic Model (BGSM) to alleviate such shortcomings. By means of maximum complete sub-graph (clique) for context selection, BGSM is capable of effectively modeling word sense representation instead of the word form itself. The proposed model is applied to phrase pair translation probability estimation and generation for SMT. The empirical results show that BGSM can enhance SMT both in performance (up to +1.3 BLEU) and efficiency in comparison against existing methods.
【Keywords】:
【Paper Link】 【Pages】:2957-2964
【Authors】: Zhuhao Wang ; Fei Wu ; Weiming Lu ; Jun Xiao ; Xi Li ; Zitong Zhang ; Yueting Zhuang
【Abstract】: Generally speaking, different persons tend to describe images from various aspects due to their individually subjective perception. As a result, generating the appropriate descriptions of images with both diversity and high quality is of great importance. In this paper, we propose a framework called GroupTalk to learn multiple image caption distributions simultaneously and effectively mimic the diversity of the image captions written by human beings. In particular, a novel iterative update strategy is proposed to separate training sentence samples into groups and learn their distributions at the same time. Furthermore, we introduce an efficient classifier to solve the problem brought about by the non-linear and discontinuous nature of language distributions which will impair performance. Experiments on several benchmark datasets show that GroupTalk naturally diversifies the generated captions of each image without sacrificing the accuracy.
【Keywords】:
【Paper Link】 【Pages】:2965-2971
【Authors】: Ruobing Xie ; Zhiyuan Liu ; Maosong Sun
【Abstract】: Representation learning of knowledge graphs aims to encode both entities and relations into a continuous low-dimensional vector space. Most existing methods only concentrate on learning representations with structured information located in triples, regardless of the rich information located in hierarchical types of entities, which could be collected in most knowledge graphs. In this paper, we propose a novel method named Type-embodied Knowledge Representation Learning (TKRL) to take advantages of hierarchical entity types. We suggest that entities should have multiple representations in different types. More specifically, we consider hierarchical types as projection matrices for entities, with two type encoders designed to model hierarchical structures. Meanwhile, type information is also utilized as relation-specific type constraints. We evaluate our models on two tasks including knowledge graph completion and triple classification, and further explore the performances on long-tail dataset. Experimental results show that our models significantly outperform all baselines on both tasks, especially with long-tail distribution. It indicates that our models are capable of capturing hierarchical type information which is significant when constructing representations of knowledge graphs. The source code of this paper can be obtained from https://github.com/thunlp/TKRL.
【Keywords】:
【Paper Link】 【Pages】:2972-2978
【Authors】: Jun Yin ; Xin Jiang ; Zhengdong Lu ; Lifeng Shang ; Hang Li ; Xiaoming Li
【Abstract】: This paper presents an end-to-end neural network model, named Neural Generative Question Answering (GENQA), that can generate answers to simple factoid questions, based on the facts in a knowledge-base. More specifically, the model is built on the encoder-decoder framework for sequence-to-sequence learning, while equipped with the ability to enquire the knowledge-base, and is trained on a corpus of question-answer pairs, with their associated triples in the knowledge-base. Empirical study shows the proposed model can effectively deal with the variations of questions and answers, and generate right and natural answers by referring to the facts in the knowledge-base. The experiment on question answering demonstrates that the proposed model can outperform an embedding-based QA model as well as a neural dialogue model trained on the same data.
【Keywords】:
【Paper Link】 【Pages】:2979-2985
【Authors】: Yichun Yin ; Furu Wei ; Li Dong ; Kaimeng Xu ; Ming Zhang ; Ming Zhou
【Abstract】: In this paper, we develop a novel approach to aspect term extraction based on unsupervised learning of distributed representations of words and dependency paths. The basic idea is to connect two words (w1 and w2) with the dependency path (r) between them in the embedding space. Specifically, our method optimizes the objective w1 + r ≈ w2 in the low-dimensional space, where the multi-hop dependency paths are treated as a sequence of grammatical relations and modeled by a recurrent neural network. Then, we design the embedding features that consider linear context and dependency context information, for the conditional random field (CRF) based aspect term extraction. Experimental results on the SemEval datasets show that, (1) with only embedding features, we can achieve state-of-the-art results; (2) our embedding method which incorporates the syntactic information among words yields better performance than other representative ones in aspect term extraction.
【Keywords】:
【Paper Link】 【Pages】:2986-2992
【Authors】: Wei Zhang ; Quan Yuan ; Jiawei Han ; Jianyong Wang
【Abstract】: We investigate the problem of personalized review-based rating prediction which aims at predicting users' ratings for items that they have not evaluated by using their historical reviews and ratings. Most of existing methods solve this problem by integrating topic model and latent factor model to learn interpretable user and items factors. However, these methods cannot utilize word local context information of reviews. Moreover, it simply restricts user and item representations equivalent to their review representations, which may bring some irrelevant information in review text and harm the accuracy of rating prediction. In this paper, we propose a novel Collaborative Multi-Level Embedding (CMLE) model to address these limitations. The main technical contribution of CMLE is to integrate word embedding model with standard matrix factorization model through a projection level.This allows CMLE to inherit the ability of capturing word local context information from word embedding model and relax the strict equivalence requirement by projecting review embedding to user and item embeddings. A joint optimization problem is formulated and solved through an efficient stochastic gradient ascent algorithm. Empirical evaluations on real datasets show CMLE outperforms several competitive methods and can solve the two limitations well.
【Keywords】:
【Paper Link】 【Pages】:2993-2999
【Authors】: Xiaodong Zhang ; Houfeng Wang
【Abstract】: Two major tasks in spoken language understanding (SLU) are intent determination (ID) and slot filling (SF). Recurrent neural networks (RNNs) have been proved effective in SF, while there is no prior work using RNNs in ID. Based on the idea that the intent and semantic slots of a sentence are correlative, we propose a joint model for both tasks. Gated recurrent unit (GRU) is used to learn the representation of each time step, by which the label of each slot is predicted. Meanwhile, a max-pooling layer is employed to capture global features of a sentence for intent classification. The representations are shared by two tasks and the model is trained by a united loss function. We conduct experiments on two datasets, and the experimental results demonstrate that our model outperforms the state-of-the-art approaches on both tasks.
【Keywords】:
【Paper Link】 【Pages】:3000-3006
【Authors】: Zhou Zhao ; Qifan Yang ; Deng Cai ; Xiaofei He ; Yueting Zhuang
【Abstract】: Expert finding for question answering is a challenging problem in Community-based Question Answering(CQA) site, arising in many applications such as question routing and the identification of best answers. In order to provide high-quality experts,many existing approaches learn the user model mainly from their past question-answering activities in CQA sites, which suffer from the sparsity problem of CQA data. In this paper, we consider the problem of expert finding from the viewpoint of learning ranking metric embedding. We propose a novel ranking metric network learning framework for expert finding by exploiting both users' relative quality rank to given questions and their social relations. We then develop a random-walk based learning method with recurrent neural networks for ranking metric network embedding. The extensive experiments on a large-scale dataset from a real world CQA site show that our method achieves better performance than other state-of-the-art solutions to the problem.
【Keywords】:
【Paper Link】 【Pages】:3007-3013
【Authors】: Xiaoqing Zheng ; Jiangtao Feng ; Mengxiao Lin ; Wenqiang Zhang
【Abstract】: Unsupervised word representations have demonstrated improvements in predictive generalization on various NLP tasks. Much effort has been devoted to effectively learning word embeddings, but little attention has been given to distributed character representations, although such character-level representations could be very useful for a variety of NLP applications in intrinsically "character-based" languages (e.g. Chinese and Japanese). On the other hand, most of existing models create a single-prototype representation per word, which is problematic because many words are in fact polysemous, and a single-prototype model is incapable of capturing phenomena of homonymy and polysemy. We present a neural network architecture to jointly learn character embeddings and induce context representations from large data sets. The explicitly produced context representations are further used to learn context-specific and multiple-prototype character embeddings, particularly capturing their polysemous variants. Our character embeddings were evaluated on three NLP tasks of character similarity, word segmentation and named entity recognition, and the experimental results demonstrated the proposed method outperformed other competing ones on all the three tasks.
【Keywords】:
【Paper Link】 【Pages】:3014-3021
【Authors】: Deyu Zhou ; Haiyang Xu ; Xin-Yu Dai ; Yulan He
【Abstract】: Storyline extraction from news streams aims to extract events under a certain news topic and reveal how those events evolve over time. It requires algorithms capable of accurately extracting events from news articles published in different time periods and linking these extracted events into coherent stories. The two tasks are often solved separately, which might suffer from the problem of error propagation. Existing unified approaches often consider events as topics, ignoring their structured representations. In this paper, we propose a non-parametric generative model to extract structured representations and evolution patterns of storylines simultaneously. In the model, each storyline is modelled as a joint distribution over some locations, organizations, persons, keywords and a set of topics. We further combine this model with the Chinese restaurant process so that the number of storylines can be determined automatically without human intervention. Moreover, per-token Metropolis-Hastings sampler based on light latent Dirichlet allocation is employed to reduce sampling complexity. The proposed model has been evaluated on three news corpora and the experimental results show that it outperforms several baseline approaches.
【Keywords】:
【Paper Link】 【Pages】:3022-3029
【Authors】: Ron Alford ; Vikas Shivashankar ; Mark Roberts ; Jeremy Frank ; David W. Aha
【Abstract】: Considerable work has focused on enhancing the semantics of Hierarchical Task Networks (HTNs) in order to advance the state-of-the-art in hierarchical planning. For instance, the Hierarchical Goal Netwwork (HGN) formalism operates over a hierarchy of goals to facilitate tighter integration of decompositional planning with classical planning. Another example is the Action Notation Markup Language (ANML) which adds aspects of generative planning and task-sharing to the standard HTN semantics.The aim of this work is to formally analyze the effects of these modifications to HTN semantics on the computational complexity and expressivity of HTN planning. To facilitate analysis, we unify goal and task planning into Goal-Task Network (GTN) planning. GTN models use HTN and HGN constructs, but have a solution-preserving mapping back to HTN planning. We then show theoretical results that provide new insights into both the expressivity as well as computational complexity of GTN planning under a number of different semantics. Our work lays a firm footing to clarify exact semantics for recent planners based on ANML, HGNs, and similar hierarchical languages.
【Keywords】:
【Paper Link】 【Pages】:3029-3039
【Authors】: Aijun Bai ; Siddharth Srivastava ; Stuart J. Russell
【Abstract】: State abstraction is an important technique for scaling MD Palgorithms. As is well known, however, it introduces difficulties due to the non-Markovian nature of state-abstracted models. Whereas prior approaches rely upon ad hoc fixes for this issue, we propose instead to view the state-abstracted model as a POMDP and show that we can thereby take advantage of state abstraction without sacrificing the Markov property. We further exploit the hierarchical structure introduced by state abstraction by extending the theory of options to a POMDP setting. In this context we propose a hierarchical Monte Carlo tree search algorithm and show that it converges to a recursively optimal hierarchical policy. Both theoretical and empirical results suggest that abstracting an MDP into a POMDP yields a scalable solution approach.
【Keywords】:
【Paper Link】 【Pages】:3038-3044
【Authors】: Arthur Bit-Monnot ; Malik Ghallab ; Félix Ingrand
【Abstract】: Planning and acting in a dynamic environment require distinguishing controllable and contingent events and checking the dynamic controllability of plans. Known procedures for testing the dynamic controllability assume that all contingent events are observable. Often this assumption does not hold. We consider here the general case of networks with invisible as well as observable contingent points. We propose a first procedure for testing their dynamic controllability. Further, we define an algorithm for choosing among the observable contingent points which to observe with additional sensing actions, such as to make a plan dynamically controllable. We show how these procedures can be incrementally integrated into a constraint-based temporal planner.
【Keywords】:
【Paper Link】 【Pages】:3045-3052
【Authors】: Blai Bonet ; Hector Geffner
【Abstract】: The problem of belief tracking in the presence of stochastic actions and observations is pervasive and yet computationally intractable. In this work we show however that probabilistic beliefs can be maintained in factored form exactly and efficiently across a number of causally closed beams, when the state variables that appear in more than one beam obey a form of backward determinism. Since computing marginals from the factors is still computationally intractable in general, and variables appearing in several beams are not always backward-deterministic, the basic formulation is extended with two approximations: forms of belief propagation for computing marginals from factors, and sampling of non-backward-deterministic variables for making such variables backward deterministic given their sampled history. Unlike, Rao-Blackwellized particle-filtering, the sampling is not used for making inference tractable but for making the factorization sound. The resulting algorithm involves sampling and belief propagation or just one of them as determined by the structure of the model.
【Keywords】:
【Paper Link】 【Pages】:3053-3059
【Authors】: Daniel Bryce ; J. Benton ; Michael W. Boldt
【Abstract】: When engineering an automated planning model, domain authors typically assume a static, unchanging ground-truth world. Unfortunately, this assumption can clash with reality, where domain changes often rapidly occur in best practices, effectors, or known conditions. In these cases, remodeling the domain causes domain experts to ensure newly captured requirements integrate well with the current model. In this work, we address this model maintenance problem in a system called Marshal. Marshal assists model maintainers by reasoning about their model as a (hidden) stochastic process. It issues queries, and learns models by observing query answers, plan solutions, and direct changes to the model. Our results indicate that anticipating model evolution leads to more accurate models over naive approaches.
【Keywords】:
【Paper Link】 【Pages】:3060-3066
【Authors】: Nicolas Catusse ; Hadrien Cambazard ; Nadia Brauner ; Pierre Lemaire ; Bernard Penz ; Anne-Marie Lagrange ; Pascal Rubini
【Abstract】: We address a parallel machine scheduling problem for which the objective is to maximize the weighted number of scheduled tasks, and with the special constraint that each task has a mandatory processing instant. This problem arises, in our case, to schedule a set of astronomical observations on a telescope. We prove that the problem is NP-complete, and we propose a constraint- programming-based branch-and-price algorithm to solve it. Experiments on real and realistic datasets show that the method provides optimal solutions very efficiently.
【Keywords】:
【Paper Link】 【Pages】:3067-3074
【Authors】: Liron Cohen ; Tansel Uras ; T. K. Satish Kumar ; Hong Xu ; Nora Ayanian ; Sven Koenig
【Abstract】: Multi-Agent Path Finding (MAPF) with the objective to minimize the sum of the travel times of the agents along their paths is a hard combinatorial problem. Recent work has shown that bounded-suboptimal MAPF solvers, such as Enhanced Conflict-Based Search or ECBS(w1) for short, run often faster than optimal MAPF solvers at the cost of incurring a suboptimality factor w1, that is due to using focal search. Other recent work has used experience graphs to guide the search of ECBS(w1) and speed it up, at the cost of incurring a separate suboptimality factor w2, that is due to inflating the heuristic values. Thus, the combination has suboptimality factor w1w2 .In this first feasibility study, we develop a bounded-suboptimal MAPF solver, Improved-ECBS or iECBS(w1) for short, that has sub optimality factor w1 rather than w1w2 (because it uses experience graphs to guide its search without inflating the heuristic values) and can run faster than ECBS(w1). We also develop two first approaches for automatically generating experience graphs for a given MAPF instance. Finally, we observe heavy-tailed behavior in the runtimes of these MAPF solvers and develop a simple rapid randomized restart strategy that can increase the success rate of iECBS(w1) within a given runtime limit.
【Keywords】:
【Paper Link】 【Pages】:3075-3081
【Authors】: Hao Cui ; Roni Khardon
【Abstract】: This paper investigates online stochastic planning for problems with large factored state and action spaces. We introduce a novel algorithm that builds a symbolic representation capturing an approximation of the action-value Q-function in terms of action variables, and then performs gradient based search to select an action for the current state. The algorithm can be seen as a symbolic extension of Monte-Carlo search, induced by independence assumptions on state and action variables, and augmented with gradients to speed up the search. This avoids the space explosion typically faced by symbolic methods, and the dearth of samples faced by Monte-Carlo methods when the action space is large. An experimental evaluation on benchmark problems shows that the algorithm is competitive with state of the art across problem sizes and that it provides significant improvements for large factored action spaces.
【Keywords】:
【Paper Link】 【Pages】:3082-3088
【Authors】: Guillem Francès ; Hector Geffner
【Abstract】: Existentially quantified variables in goals and action preconditions are part of the standard PDDL planning language, yet few planners support them, while those that do compile them away at an exponential cost. In this work, we argue that existential variables are an essential feature for representing and reasoning with constraints in planning, and that it is harmful to compile them away or avoid them altogether, since this hides part of the problem structure that can be exploited computationally. We show how to do this by formulating an extension of the standard delete-relaxation heuristics that handles existential variables. While this extension is simple, the consequences for both modeling and computation are important. Furthermore, by allowing existential variables in STRIPS and treating them properly, CSPs can be represented and solved in a direct manner as action-less,fluent-less STRIPS planning problems, something important for problems involving restrictions. In addition, functional fluents in Functional STRIPS can be compiled away with no effect on the structure and informativeness of the resulting heuristic. Experiments are reported comparing our native ∃-STRIPS planner with state-of-the-art STRIPS planners over compiled and propositional encodings, and with a Functional STRIPS planner.
【Keywords】:
【Paper Link】 【Pages】:3089-3095
【Authors】: Caelan Reed Garrett ; Leslie Pack Kaelbling ; Tomás Lozano-Pérez
【Abstract】: We investigate learning heuristics for domain-specific planning. Prior work framed learning a heuristic as an ordinary regression problem. However, in a greedy best-first search, the ordering of states induced by a heuristic is more indicative of the resulting planner's performance than mean squared error. Thus, we instead frame learning a heuristic as a learning to rank problem which we solve using a RankSVM formulation. Additionally, we introduce new methods for computing features that capture temporal interactions in an approximate plan. Our experiments on recent International Planning Competition problems show that the RankSVM learned heuristics outperform both the original heuristics and heuristics learned through ordinary regression.
【Keywords】:
【Paper Link】 【Pages】:3096-3102
【Authors】: Supriyo Ghosh ; Michael Trick ; Pradeep Varakantham
【Abstract】: Bike Sharing Systems (BSSs) experience a significant loss in customer demand due to starvation (empty base stations precluding bike pickup) or congestion (full base stations precluding bike return). Therefore, BSSs operators reposition bikes between stations with the help of carrier vehicles. Due to unpredictable and dynamically changing nature of the demand, myopic reasoning typically provides a below par performance. We propose an online and robust repositioning approach to minimise the loss in customer demand while considering the possible uncertainty in future demand. Specifically, we develop a scenario generation approach based on an iterative two player game to compute a strategy of repositioning by assuming that the environment can generate a worse demand scenario (out of the feasible demand scenarios) against the current repositioning solution. Extensive computational results from a simulation built on real world data set of bike sharing company demonstrate that our approach can significantly reduce the expected lost demand over the existing benchmark approaches.
【Keywords】:
【Paper Link】 【Pages】:3103-3109
【Authors】: Vincent Gingras ; Claude-Guy Quimper
【Abstract】: We present two novel filtering algorithms for the Cumulative constraint based on a new energetic relaxation. We introduce a generalization of the Overload Check and Edge-Finder rules based on a function computing the earliest completion time for a set of tasks. Depending on the relaxation used to compute this function, one obtains different levels of filtering. We present two algorithms that enforce these rules. The algorithms utilize a novel data structure that we call Profile and that encodes the resource utilization over time. Experiments show that these algorithms are competitive with the state-of-the-art algorithms, by doing a greater filtering and having a faster runtime.
【Keywords】:
【Paper Link】 【Pages】:3110-3116
【Authors】: Daniel Gnad ; Martin Wehrle ; Jörg Hoffmann
【Abstract】: Recent work has introduced fork-decoupled search, addressing classical planning problems where a single center component provides preconditions for several leaf components. Given a fixed center path πC, the leaf moves compliant with πC can then be scheduled independently for each leaf. Fork-decoupled search thus searches over center paths only, maintaining the compliant paths for each leaf separately. This can yield dramatic benefits. It is empirically complementary to partial order reduction via strong stubborn sets, in that each method yields its strongest reductions in different benchmarks. Here we show that the two methods can be combined, in the form of strong stubborn sets for fork-decoupled search. This can yield exponential advantages relative to both methods. Empirically, the combination reliably inherits the best of its components, and often outperforms both.
【Keywords】:
【Paper Link】 【Pages】:3117-3123
【Authors】: Ragavendran Gopalakrishnan ; Arpita Biswas ; Alefiya Lightwala ; Skanda Vasudevan ; Partha Dutta ; Abhishek Tripathi
【Abstract】: Effective placement of charging stations plays a key role in Electric Vehicle (EV) adoption. In the placement problem, given a set of candidate sites, an optimal subset needs to be selected with respect to the concerns of both (a) the charging station service provider, such as the demand at the candidate sites and the budget for deployment, and (b) the EV user, such as charging station reachability and short waiting times at the station. This work addresses these concerns, making the following three novel contributions: (i) a supervised multi-view learning framework using Canonical Correlation Analysis (CCA) for demand prediction at candidate sites, using multiple datasets such as points of interest information, traffic density, and the historical usage at existing charging stations; (ii) a mixed-packing-and-covering optimization framework that models competing concerns of the service provider and EV users; (iii) an iterative heuristic to solve these problems by alternately invoking knapsack and setcover algorithms. The performance of the demand prediction model and the placement optimization heuristic are evaluated using real world data.
【Keywords】:
【Paper Link】 【Pages】:3124-3132
【Authors】: Eric A. Hansen ; Jinchuan Shi ; Arindam Khaled
【Abstract】: We propose a node-removal/arc-reversal algorithm for influence diagram evaluation that includes reductions that allow an influence diagram to be solved by a generalization of the dynamic programming approach to solving partially observable Markov decision processes (POMDPs). Among its potential advantages, the algorithm allows a more flexible ordering of node removals, and a POMDP-inspired approach to optimizing over hidden state variables, which can improve the scalability of influence diagram evaluation in solving complex, multi-stage problems. It also finds a more compact representation of an optimal strategy.
【Keywords】:
【Paper Link】 【Pages】:3133-3139
【Authors】: Netantel Hasidi ; Roni Stern ; Meir Kalech ; Shulamit Reches
【Abstract】: Troubleshooting is the process of diagnosing and repairing a system that is behaving abnormally. Diagnostic and repair actions may incur costs, and traditional troubleshooting algorithms aim to minimize the costs incurred until the system is fixed. We propose an anticipatory troubleshooting algorithm, which is able to reason about both current and future failures. To reason about failures over time, we incorporate statistical tools from survival analysis that enable predicting when a failure is likely to occur. Incorporating this prognostic information in a troubleshooting algorithm enables (1) better fault isolation and (2) more intelligent decision making in which repair actions to employ to minimize troubleshooting costs over time.
【Keywords】:
【Paper Link】 【Pages】:3140-3146
【Authors】: Chao Huang ; Xin Chen ; Yifan Zhang ; Shengchao Qin ; Yifeng Zeng ; Xuandong Li
【Abstract】: Ensuring the stability is the most important requirement for the navigation control of multi-robot systems with no reference trajectory. The popular heuristic-search methods cannot provide theoretical guarantees on stability. In this paper, we propose a Hierarchical Model Predictive Control scheme that employs reachable sets to decouple the navigation problem of linear dynamical multi-robot systems. The proposed control scheme guarantees the stability and feasibility, and is more efficient and viable than other Model Predictive Control schemes, as evidenced by our simulation results.
【Keywords】:
【Paper Link】 【Pages】:3147-3153
【Authors】: Shivaram Kalyanakrishnan ; Utkarsh Mall ; Ritish Goyal
【Abstract】: Policy Iteration (PI) is a widely-used family of algorithms for computing an optimal policy for a given Markov Decision Problem (MDP). Starting with an arbitrary initial policy, PI repeatedly updates to a dominating policy until an optimal policy is found. The update step involves switching the actions corresponding to a set of "improvable" states, which are easily identified. Whereas progress is guaranteed even if just one improvable state is switched at every step, the canonical variant of PI, attributed to Howard [1960], switches every improvable state in order to obtain the next iterate. For MDPs with n states and 2 actions per state, the tightest known bound on the complexity of Howard's PI is O(2n/n) iterations. To date, the tightest bound known across all variants of PI is O(1.7172n) expected iterations for a randomised variant introduced by Mansour and Singh [1999]. We introduce Batch-Switching Policy Iteration (BSPI), a family of deterministic PI algorithms that switches states in "batches," taking the batch size b as a parameter. By varying b, BSPI interpolates between Howard's PI and another previously-studied variant called Simple PI [Melekopoglou and Condon, 1994]. Our main contribution is a bound of O(1.6479n) on the number of iterations taken by an instance of BSPI. We believe this is the tightest bound shown yet for any variant of PI. We also present experimental results that suggest Howard's PI might itself enjoy an even tighter bound.
【Keywords】:
【Paper Link】 【Pages】:3154-3160
【Authors】: Michael Katz ; Vitaly Mirkis
【Abstract】: The objective of partial satisfaction planning is to achieve an as valuable as possible state, tacking into account the cost of its achievement. In this work we investigate the computational complexity of restricted fragments of two variants of partial satisfaction: net-benefit and oversubscription planning. In particular, we examine restrictions on the causal graph structure and variable domain size of the planning problem, and show that even for the strictest such restrictions, optimal oversubscription planning is hard. In contrast, certain tractability results previously obtained for classical planning also apply to net-benefit planning. We then partially relax these restrictions in order to find the boundary of tractability for both variants of partial satisfaction planning. In addition, for the family of $0$-binary value functions we show a strong connection between the complexity of cost-optimal classical and optimal oversubscription planning.
【Keywords】:
【Paper Link】 【Pages】:3161-3169
【Authors】: Thomas Keller ; Florian Pommerening ; Jendrik Seipp ; Florian Geißer ; Robert Mattmüller
【Abstract】: Abstraction heuristics are a popular method to guide optimal search algorithms in classical planning. Cost partitionings allow to sum heuristic estimates admissibly by distributing action costs among the heuristics. We introduce state-dependent cost partitionings which take context information of actions into account, and show that an optimal state-dependent cost partitioning dominates its state-independent counterpart. We demonstrate the potential of our idea with a state-dependent variant of the recently proposed saturated cost partitioning, and show that it has the potential to improve not only over its state-independent counterpart, but even over the optimal state-independent cost partitioning. Our empirical results give evidence that ignoring the context of actions in the computation of a cost partitioning leads to a significant loss of information.
【Keywords】:
【Paper Link】 【Pages】:3170-3176
【Authors】: Sarah Keren ; Avigdor Gal ; Erez Karpas
【Abstract】: Big brother is watching but his eyesight is not all that great, since he only has partial observability of the environment. In such a setting agents maybe able to preserve their privacy by hiding their true goal, following paths that may lead to multiple goals. In this work we present a framework that supports the offline analysis of goal recognition settings with non-deterministic system sensor models, in which the observer has partial (and possibly noisy) observability of the agent's actions, while the agent is assumed to have full observability of his environment. In particular, we propose anew variation of worst case distinctiveness (wcd), a measure that assesses the ability to perform goal recognition within a model. We describe a new, efficient way to compute this measure via a novel compilation to classical planning. In addition, we discuss the tools agents have to preserve privacy, by keeping their goal ambiguous as long as possible. Our empirical evaluation shows the feasibility of the proposed solution.
【Keywords】:
【Paper Link】 【Pages】:3177-3184
【Authors】: Christian Kroer ; Tuomas Sandholm
【Abstract】: Biological adaptation is a powerful mechanism that makes many disorders hard to combat. In this paper we study steering such adaptation through sequential planning. We propose a general approach where we leverage Monte Carlo tree search to compute a treatment plan, and the biological entity is modeled by a black-box simulator that the planner calls during planning. We show that the framework can be used to steer a biological entity modeled via a complex signaling pathway network that has numerous feedback loops that operate at different rates and have hard-to-understand aggregate behavior. We apply the framework to steering the adaptation of a patient's immune system. In particular, we apply it to a leading T cell simulator (available in the biological modeling package BioNetGen). We run experiments with two alternate goals: developing regulatory T cells or developing effector T cells. The former is important for preventing autoimmune diseases while the latter is associated with better survival rates in cancer patients. We are especially interested in the effect of sequential plans, an approach that has not been explored extensively in the biological literature. We show that for the development of regulatory cells, sequential plans yield significantly higher utility than the best static therapy. In contrast, for developing effector cells, we find that (at least for the given simulator, objective function, action possibilities, and measurement possibilities) single-step plans suffice for optimal treatment.
【Keywords】:
【Paper Link】 【Pages】:3185-3191
【Authors】: Levi H. S. Lelis ; Santiago Franco ; Marvin Abisrror ; Mike Barley ; Sandra Zilles ; Robert C. Holte
【Abstract】: In this paper we present greedy methods for selecting a subset of heuristic functions for guiding A search. Our methods are able to optimize various objective functions while selecting a subset from a pool of up to thousands of heuristics. Specifically, our methods minimize approximations of A's search tree size, and approximations of A*'s running time. We show empirically that our methods can outperform state-of-the-art planners for deterministic optimal planning.
【Keywords】:
【Paper Link】 【Pages】:3192-3198
【Authors】: Yunlong Liu ; Hexing Zhu ; Yifeng Zeng ; Zongxiong Dai
【Abstract】: Predictive State Representations (PSRs) are an efficient method for modelling partially observable dynamical systems. They have shown advantages over the latent state-based approaches by using functions of a set of observable quantities, called tests, to represent model states. As a consequence, discovering the set of tests for representing the state is one of the central problems in PSRs. Existing techniques either discover these tests through iterative methods, which can only be applied to some toy problems, or avoid the complex discovery problem by maintaining a very large set of tests, which may be prohibitively expensive. In this paper, with the benefits of Monte-Carlo tree search (MCTS) for finding solutions in complex problems, and by proposing the concept of model entropy for measuring the model accuracy as the evaluation function in MCTS, the discovery problem is formalized as a sequential decision making problem. Then such a decision making problem can be solved using MCTS, a set of tests for representing state can be obtained and the PSR model of the underlying system can be learned straightforwardly. We conduct experiments on several domains including one extremely large domain and the experimental results show the effectiveness of our approach.
【Keywords】:
【Paper Link】 【Pages】:3199-3205
【Authors】: Damir Lotinac ; Javier Segovia Aguas ; Sergio Jiménez Celorrio ; Anders Jonsson
【Abstract】: In many domains generalized plans can only be computed if certain high-level state features, i.e. ~features that capture key concepts to accurately distinguish between states and make good decisions, are available. In most applications of generalized planning such features are hand-coded by an expert. This paper presents a novel method to automatically generate high-level state features for solving a generalized planning problem. Our method extends a compilation of generalized planning into classical planning and integrates the computation of generalized plans with the computation of features, in the form of conjunctive queries. Experiments show that we generate features for diverse generalized planning problems and hence, compute generalized plans without providing a prior high-level representation of the states. We also bring a new landscape of challenging benchmarks to classical planning since our compilation naturally models classification tasks as classical planning problems.
【Keywords】:
【Paper Link】 【Pages】:3206-3212
【Authors】: Christian J. Muise ; Paolo Felli ; Tim Miller ; Adrian R. Pearce ; Liz Sonenberg
【Abstract】: Single-agent planning in a multi-agent environment is challenging because the actions of other agents can affect our ability to achieve a goal. From a given agent's perspective, actions of others can be viewed as non-deterministic outcomes of that agent's actions. While simple conceptually, this interpretation of planning in a multi-agent environment as non-deterministic planning remains challenging, not only due to the non-determinism resulting from others' actions, but because it is not clear how to compactly model the possible actions of others in the environment. In this paper, we cast the problem of planning in a multi-agent environment as one of Fully-Observable Non-Deterministic (FOND) planning. We extend a non-deterministic planner to plan in a multi-agent setting, allowing non-deterministic planning technology to solve a new class of planning problems. To improve the efficiency in domains too large for solving optimally, we propose a technique to use the goals and possible actions of other agents to focus the search on a set of plausible actions. We evaluate our approach on existing and new multi-agent benchmarks, demonstrating that modelling the other agents' goals improves the quality of the resulting solutions.
【Keywords】:
【Paper Link】 【Pages】:3213-3219
【Authors】: Wiktor Mateusz Piotrowski ; Maria Fox ; Derek Long ; Daniele Magazzeni ; Fabio Mercorio
【Abstract】: Planning with hybrid domains modelled in PDDL+ has been gaining research interest in the Automated Planning community in recent years. Hybrid domain models capture a more accurate representation of real world problems that involve continuous processes than is possible using discrete systems. However, solving problems represented as PDDL+ domains is very challenging due to the construction of complex system dynamics, including non-linear processes and events. In this paper we introduce DiNo, a new planner capable of tackling complex problems with non-linear system dynamcs governing the continuous evolution of states. DiNo is based on the discretise-and-validate approach and uses the novel Staged Relaxed Planning Graph+ (SRPG+) heuristic, which is introduced in this paper. Although several planners have been developed to work with subsets of PDDL+ features, or restricted forms of processes, DiNo is currently the only heuristic planner capable of handling non-linear system dynamics combined with the full PDDL+ feature set.
【Keywords】:
【Paper Link】 【Pages】:3220-3227
【Authors】: Yash Satsangi ; Shimon Whiteson ; Frans A. Oliehoek
【Abstract】: Submodular function maximization finds application in a variety of real-world decision-making problems. However, most existing methods, based on greedy maximization, assume it is computationally feasible to evaluate F, the function being maximized. Unfortunately, in many realistic settings F is too expensive to evaluate exactly even once. We present probably approximately correct greedy maximization, which requires access only to cheap anytime confidence bounds on F and uses them to prune elements. We show that, with high probability, our method returns an approximately optimal set. We also propose novel, cheap confidence bounds for conditional entropy, which appears in many common choices of F and for which it is difficult to find unbiased or bounded estimates. Finally, results on a real-world dataset from a multi-camera tracking system in a shopping mall demonstrate that our approach performs comparably to existing methods, but at a fraction of the computational cost.
【Keywords】:
【Paper Link】 【Pages】:3228-3234
【Authors】: Enrico Scala ; Patrik Haslum ; Sylvie Thiébaux
【Abstract】: The paper presents a new relaxation for hybrid planning with continuous numeric and propositional state variables based on subgoaling, generalising the subgoaling principle underlying the hmax and hadd heuristics to such problems. Our relaxation improves on existing interval-based relaxations by taking into account some negative interactions between effects when achieving a subgoal, resulting in better estimates. We show conditions on the planning model ensuring that this new relaxation leads to tractable, and, for the hmax version, admissible, heuristics. The new relaxation can be combined with the interval-based relaxation, to derive heuristics applicable to general numeric planning, while still providing more informed estimates for the subgoals that meet these conditions. Experiments show the effectiveness of its inadmissible and admissible version on satisficing and optimal numeric planning,respectively. As far as we know, this is the first admissible heuristic enabling cost-optimal numeric planning.
【Keywords】:
【Paper Link】 【Pages】:3235-3241
【Authors】: Javier Segovia Aguas ; Sergio Jiménez Celorrio ; Anders Jonsson
【Abstract】: Finite State Controllers (FSCs) are an effective way to represent sequential plans compactly. By imposing appropriate conditions on transitions, FSCs can also represent generalized plans that solve a range of planning problems from a given domain. In this paper we introduce the concept of hierarchical FSCs for planning by allowing controllers to call other controllers. We show that hierarchical FSCs can represent generalized plans more compactly than individual FSCs. Moreover, our call mechanism makes it possible to generate hierarchical FSCs in a modular fashion, or even to apply recursion. We also introduce a compilation that enables a classical planner to generate hierarchical FSCs that solve challenging generalized planning problems. The compilation takes as input a set of planning problems from a given domain and outputs a single classical planning problem, whose solution corresponds to a hierarchical FSC.
【Keywords】:
【Paper Link】 【Pages】:3242-3250
【Authors】: Jendrik Seipp ; Florian Pommerening ; Gabriele Röger ; Malte Helmert
【Abstract】: We analyze how complex a heuristic function must be to directly guide a state-space search algorithm towards the goal. As a case study, we examine functions that evaluate states with a weighted sum of state features. We measure the complexity of a domain by the complexity of the required features. We analyze conditions under which the search algorithm runs in polynomial time and show complexity results for several classical planning domains.
【Keywords】:
【Paper Link】 【Pages】:3251-3257
【Authors】: Alexander Shleyfman ; Alexander Tuisov ; Carmel Domshlak
【Abstract】: Similarly to the classical AI planning, the Atari 2600 games supported in the Arcade Learning Environment all feature a fully observable (RAM) state and actions that have deterministic effect. At the same time, the problems in ALE are given only implicitly, via a simulator, a priori precluding exploiting most of the modern classical planning techniques. Despite that, Lipovetzky et al. [2015] recently showed how online planning for Atari-like problems can be effectively addressed using IW(i), a blind state-space search algorithm that employs a certain form of structural similarity-based pruning. We show that the effectiveness of the blind state-space search for Atari-like online planning can be pushed even further by focusing the search using both structural state similarity and the relative myopic value of the states. We also show that the planning effectiveness can be further improved by considering online planning for the Atari games as a multiarmed bandit style competition between the various actions available at the state planned for, and not purely as a classical planning style action sequence optimization problem.
【Keywords】:
【Paper Link】 【Pages】:3258-3264
【Authors】: Shirin Sohrabi ; Anton V. Riabov ; Octavian Udrea
【Abstract】: In this paper, we propose to extend previous work to (1) address observations over fluents, (2) better address unreliable observations (i.e., noisy or missing observations), and (3) recognize plans in addition to goals. To this end, we introduce a relaxation of the plan-recognition-as-planning formulation that allows unreliable observations. That is, in addition to the original costs of the plan, we define two objectives that account for missing and noisy observations, and optimize for a linear combination of all objectives. We approximate the posterior probabilities of generated plans by taking into account the combined costs that include penalties for missing or noisy observations, and normalizing over a sample set of plans generated by finding either diverse or high-quality plans. Our experiments show that this approach improves goal recognition in most domains when observations are unreliable. In addition, we evaluate plan recognition performance and show that the high-quality plan generation approach should be preferred in most domains.
【Keywords】:
【Paper Link】 【Pages】:3265-3271
【Authors】: Álvaro Torralba ; Daniel Gnad ; Patrick Dubbert ; Jörg Hoffmann
【Abstract】: Fork-decoupled search is a recent approach to classical planning that exploits fork structures, where a single center component provides preconditions for several leaf components. The decoupled states in this search consist of a center state, along with a price for every leaf state. Given this, when does one decoupled state dominate another? Such state-dominance criteria can be used to prune dominated search states. Prior work has devised only a trivial criterion. We devise several more powerful criteria, show that they preserve optimality, and establish their interrelations. We show that they can yield exponential reductions. Experiments on IPC benchmarks attest to the possible practical benefits.
【Keywords】:
【Paper Link】 【Pages】:3272-3278
【Authors】: Álvaro Torralba ; Carlos Linares López ; Daniel Borrajo
【Abstract】: Symbolic bidirectional uniform-cost search is a prominent technique for cost-optimal planning. Thus, the question whether it can be further improved by making use of heuristic functions raises naturally. However, the use of heuristics in bidirectional search does not always improve its performance. We propose a novel way to use abstraction heuristics in symbolic bidirectional search in which the search only resorts to heuristics when it becomes unfeasible. We adapt the definition of partial and perimeter abstractions to bidirectional search, where A is used to traverse the abstract state spaces and/or generate the perimeter. The results show that abstraction heuristics can further improve symbolic bidirectional search in some domains. In fact, the resulting planner, SymBA, was the winner of the optimal-track of the last IPC.
【Keywords】:
【Paper Link】 【Pages】:3279-3285
【Authors】: Christabel Wayllace ; Ping Hou ; William Yeoh ; Tran Cao Son
【Abstract】: Goal Recognition Design (GRD) problems involve identifying the best ways to modify the underlying environment that the agents operate in, typically by making a subset of feasible actions infeasible, in such a way that agents are forced to reveal their goals as early as possible. Thus far, existing work assumes that the outcomes of the actions of the agents are deterministic, which might be unrealistic in real-world problems. For example, wheel slippage in robots cause the outcomes of their movements to be stochastic. In this paper, we generalize the GRD problem to Stochastic GRD (S-GRD) problems, which handle stochastic action outcomes. We also generalize the worst-case distinctiveness (wcd) measure, which measures the goodness of a solution, to take stochasticity into account. Finally, we introduce Markov decision process (MDP) based algorithms to compute the wcd and minimize it by making up to k actions infeasible.
【Keywords】:
【Paper Link】 【Pages】:3286-3292
【Authors】: Martin Wehrle ; Silvan Sievers ; Malte Helmert
【Abstract】: In domain-independent planning, dependencies of operators and variables often prevent the effective application of planning techniques that rely on loosely coupled problems (like factored planning or partial order reduction). In this paper, we propose a generic approach for factorizing a classical planning problem into an equivalent problem with fewer operator and variable dependencies. Our approach is based on variable factorization, which can be reduced to the well-studied problem of graph factorization. While the state spaces of the original and the factorized problems are isomorphic, the factorization offers the potential to exponentially reduce the complexity of planning techniques like factored planning and partial order reduction.
【Keywords】:
【Paper Link】 【Pages】:3293-3299
【Authors】: Dominik Winterer ; Martin Wehrle ; Michael Katz
【Abstract】: Symmetry reduction has significantly contributed to the success of classical planning as heuristic search. However, it is an open question if symmetry reduction techniques can be lifted to fully observable nondeterministic (FOND) planning. We generalize the concepts of structural symmetries and symmetry reduction to FOND planning and specifically to the LAO algorithm. Our base implementation of LAO in the Fast Downward planner is competitive with the LAO-based FOND planner myND. Our experiments further show that symmetry reduction can yield strong performance gains compared to our base implementation of LAO.
【Keywords】:
【Paper Link】 【Pages】:3300-3307
【Authors】: Peng Yu ; Jiaying Shen ; Peter Z. Yeh ; Brian Charles Williams
【Abstract】: In recent literature, several approaches have been developed to solve over-constrained travel planning problems, which are often framed as conditional temporal problems with discrete choices. These approaches are able to explain the causes of failure and recommend alternative solutions by suspending or weakening temporal constraints. While helpful, they may not be practical in many situations, as we often cannot compromise on time. In this paper, we present an approach for solving such over-constrained problems, by also relaxing non-temporal variable domains through the consideration of additional options that are semantically similar. Our solution, called Conflict-Directed Semantic Relaxation (CDSR), integrates a knowledge base and a semantic similarity calculator, and is able to simultaneously enumerate both temporal and domain relaxations in best-first order. When evaluated empirically on a range of urban trip planning scenarios, CDSR demonstrates a substantial improvement in flexibility compared to temporal relaxation only approaches.
【Keywords】:
【Paper Link】 【Pages】:3308-3314
【Authors】: Chongjie Zhang ; Julie A. Shah
【Abstract】: To enable large-scale multi-agent coordination under temporal and spatial constraints, we formulate it as a multi-level optimization problem and develop a multi-abstraction search approach for co-optimizing agent placement with task assignment and scheduling. This approach begins with a highly abstract agent placement problem and the rapid computation of an initial solution, which is then improved upon using a hill climbing algorithm for a less abstract problem; finally, the solution is fine-tuned within the original problem space. Empirical results demonstrate that this multi-abstraction approach significantly outperforms a conventional hill climbing algorithm and an approximate mixed-integer linear programming approach.
【Keywords】:
【Paper Link】 【Pages】:3315-3323
【Authors】: Qi Zhang ; Edmund H. Durfee ; Satinder P. Singh ; Anna Chen ; Stefan J. Witwicki
【Abstract】: Cooperating agents can make commitments to help each other, but commitments might have to be probabilistic when actions have stochastic outcomes. We consider the additional complication in cases where an agent might prefer to change its policy as it learns more about its reward function from experience. How should such an agent be allowed to change its policy while still faithfully pursuing its commitment in a principled decision-theoretic manner? We address this question by defining a class of Dec-POMDPs with Bayesian reward uncertainty, and by developing a novel Commitment Constrained Iterative Mean Reward algorithm that implements the semantics of faithful commitment pursuit while still permitting the agent's response to the evolving understanding of its rewards. We bound the performance of our algorithm theoretically, and evaluate empirically how it effectively balances solution quality and computation cost.
【Keywords】:
【Paper Link】 【Pages】:3324-3330
【Authors】: Congqi Cao ; Yifan Zhang ; Chunjie Zhang ; Hanqing Lu
【Abstract】: Torso joints can be considered as the landmarks of human body. An action consists of a series of body poses which are determined by the positions of the joints. With the rapid development of RGB-D camera technique and pose estimation research, the acquisition of the body joints has become much easier than before. Thus, we propose to incorporate joint positions with currently popular deep-learned features for action recognition. In this paper, we present a simple, yet effective method to aggregate convolutional activations of a 3D deep convolutional neural network (3D CNN) into discriminative descriptors based on joint positions. Two pooling schemes for mapping body joints into convolutional feature maps are discussed. The joints-pooled 3D deep convolutional descriptors (JDDs) are more effective and robust than the original 3D CNN features and other competing features. We evaluate the proposed descriptors on recognizing both short actions and complex activities. Experimental results on real-world datasets show that our method generates promising results, outperforming state-of-the-art results significantly.
【Keywords】:
【Paper Link】 【Pages】:3331-3337
【Authors】: Chen Chen ; Mengyuan Liu ; Baochang Zhang ; Jungong Han ; Junjun Jiang ; Hong Liu
【Abstract】: This paper presents an effective local spatio-temporal descriptor for action recognition from depth video sequences. The unique property of our descriptor is that it takes the shape discrimination and action speed variations into account, intending to solve the problems of distinguishing different pose shapes and identifying the actions with different speeds in one goal. The entire algorithm is carried out in three stages. In the first stage, a depth sequence is divided into temporally overlapping depth segments which are used to generate three depth motion maps (DMMs), capturing the shape and motion cues. To cope with speed variations in actions, multiple frame lengths of depth segments are utilized, leading to a multi-temporal DMMs representation. In the second stage, all the DMMs are first partitioned into dense patches. Then, the local binary patterns (LBP) descriptor is exploited to characterize local rotation invariant texture information in those patches. In the third stage, the Fisher kernel is employed to encode the patch descriptors for a compact feature representation, which is fed into a kernel-based extreme learning machine classifier. Extensive experiments on the public MSRAction3D, MSRGesture3D and DHA datasets show that our proposed method outperforms state-of-the-art approaches for depth-based action recognition.
【Keywords】:
【Paper Link】 【Pages】:3338-3344
【Authors】: Lin Chen ; Baoxin Li
【Abstract】: Semantic attributes have been proposed to bridge the semantic gap between low-level feature representation and high-level semantic understanding of visual objects. Obtaining a good representation of semantic attributes usually requires learning from high-dimensional low-level features, which not only significantly increases the time and space requirement but also degrades the performance due to numerous irrelevant features. Since multi-attribute prediction can be generalized as a multi-task learning problem, sparse-based multi-task feature selection approaches have been introduced, utilizing the relatedness among multiple attributes. However, such approaches either do not investigate the pattern of the relatedness among attributes, or require prior knowledge about the pattern. In this paper, we propose a novel feature selection approach which embeds attribute correlation modeling in multi-attribute joint feature selection. Experiments on both synthetic dataset and multiple public benchmark datasets demonstrate that the proposed approach effectively captures the correlation among multiple attributes and significantly outperforms the state-of-the-art approaches.
【Keywords】:
【Paper Link】 【Pages】:3345-3351
【Authors】: Yanhua Cheng ; Xin Zhao ; Rui Cai ; Zhiwei Li ; Kaiqi Huang ; Yong Rui
【Abstract】: This paper studies the problem of RGB-D object recognition. Inspired by the great success of deep convolutional neural networks (DCNN) in AI, researchers have tried to apply it to improve the performance of RGB-D object recognition. However, DCNN always requires a large-scale annotated dataset to supervise its training. Manually labeling such a large RGB-D dataset is expensive and time consuming, which prevents DCNN from quickly promoting this research area. To address this problem, we propose a semi-supervised multimodal deep learning framework to train DCNN effectively based on very limited labeled data and massive unlabeled data. The core of our framework is a novel diversity preserving co-training algorithm, which can successfully guide DCNN to learn from the unlabeled RGB-D data by making full use of the complementary cues of the RGB and depth data in object representation. Experiments on the benchmark RGB-D dataset demonstrate that, with only 5% labeled training data, our approach achieves competitive performance for object recognition compared with those state-of-the-art results reported by fully-supervised methods.
【Keywords】:
【Paper Link】 【Pages】:3352-3358
【Authors】: Gary Doran ; David R. Thompson ; Tara A. Estlin
【Abstract】: A key component of Mars exploration is the operation of robotic instruments on the surface, such as those on board the Mars Exploration Rovers, the Mars Science Laboratory (MSL), and the planned Mars 2020 Rover. As the instruments carried by these rovers have become more advanced, the area targeted by some instruments becomes smaller, revealing more fine-grained details about the geology and chemistry of rocks on the surface. However, thermal fluctuations, rover settling or slipping, and inherent inaccuracies in pointing mechanisms all lead to pointing error that is on the order of the target size (several millimeters) or larger. We show that given a target located on a previously acquired image, the rover can align this with a new image to visually locate the target and refine the current pointing. Due to round-trip communication constraints, this visual targeting must be done efficiently on board the rover using relatively limited computing hardware. We employ existing ORB features for landmark-based image registration, describe and theoretically justify a novel approach to filtering false landmark matches, and employ a random forest classifier to automatically reject failed alignments. We demonstrate the efficacy of our approach using over 3,800 images acquired by Remote Micro-Imager on board the Curiosity rover.
【Keywords】:
【Paper Link】 【Pages】:3359-3367
【Authors】: Mohammed E. Fathy ; Azadeh Alavi ; Rama Chellappa
【Abstract】: With the abundance of video data, the interest in more effective methods for recognizing faces from unconstrained videos has grown. State-of-the-art algorithms for describing an image set use descriptors that are either very high-dimensional and/or sensitive to outliers and image misalignment. In this paper, we represent image sets as dictionaries of Symmetric Positive Definite (SPD) matrices that are more robust to local deformations and outliers. We then learn a tangent map for transforming the SPD matrix logarithms into a lower-dimensional Log-Euclidean space such that the transformed gallery atoms adhere to a more discriminative subspace structure. A query image set is then classified by first mapping its SPD descriptors into the computed Log-Euclidean tangent space and using the sparse representation over the tangent space to decide a label for the image set. Experiments on three public video datasets show that the proposed method outperforms many state-of-the-art methods.
【Keywords】:
【Paper Link】 【Pages】:3368-3374
【Authors】: Reza Shoja Ghiass ; Ognjen Arandjelovic
【Abstract】: Determining the direction in which a person is looking is an important problem in a wide range of HCI applications. In this paper we describe a highly accurate algorithm that performs gaze estimation using an affordable and widely available device such as Kinect. The method we propose starts by performing accurate head pose estimation achieved by fitting a person specific morphable model of the face to depth data. The ordinarily competing requirements of high accuracy and high speed are met concurrently by formulating the fitting objective function as a combination of terms which excel either in accurate or fast fitting, and then by adaptively adjusting their relative contributions throughout fitting. Following pose estimation, pose normalization is done by re-rendering the fitted model as a frontal face. Finally gaze estimates are obtained through regression from the appearance of the eyes in synthetic, normalized images. Using EYEDIAP, the standard public dataset for the evaluation of gaze estimation algorithms from RGB-D data, we demonstrate that our method greatly outperforms the state of the art.
【Keywords】:
【Paper Link】 【Pages】:3375-3381
【Authors】: Jasmin Grosinger ; Federico Pecora ; Alessandro Saffiotti
【Abstract】: In order to be proactive, robots should be capable of generating and selecting their own goals, and pursuing activities towards their achievement. Goal reasoning has focused on the former set of cognitive abilities, and automated planning on the latter. Despite the existence of robots that possess both capabilities, we lack a general understanding of how to combine goal generation and goal achievement. This paper introduces the notion of equilibrium maintenance as a contribution to this understanding. We provide formal evidence that equilibrium maintenance is conducive to proactive robots, and demonstrate our approach in a closed loop with a real robot in a smart home.
【Keywords】:
【Paper Link】 【Pages】:3382-3388
【Authors】: Yuchen Guo ; Guiguang Ding ; Jungong Han ; Xiaoming Jin
【Abstract】: Iterative Quantization (ITQ) is one of the most successful hashing based nearest-neighbor search methods for large-scale information retrieval in the past a few years due to its simplicity and superior performance. However, the performance of this algorithm degrades significantly when dealing with noisy data. Additionally, it can barely facilitate a wide range of applications as the distortion measurement only limits to ℓ2 norm. In this paper, we propose an ITQ+ algorithm, aiming to enhance both robustness and generalization of the original ITQ algorithm. Specifically, a ℓp,q-norm loss function is proposed to conduct the ℓp-norm similarity search, rather than a ℓ2} norm search. Despite the fact that changing the loss function to ℓp,q-norm makes our algorithm more robust and generic, it brings us a challenge that minimizes the obtained orthogonality constrained ℓp,q-norm function, which is non-smooth and non-convex. To solve this problem, we propose a novel and efficient optimization scheme. Extensive experiments on benchmark datasets demonstrate that ITQ+ is overwhelmingly better than the original ITQ algorithm, especially when searching similarity in noisy data.
【Keywords】:
【Paper Link】 【Pages】:3389-3395
【Authors】: Allam S. Hassanein ; Mohamed E. Hussein ; Walid Gomaa
【Abstract】: In this paper we address the problem of semantic analysis of structured / unstructured crowded video scenes. Our proposed approach relies on tracklets for motion representation. Each extracted tracklet is abstracted as a directed line segment, and a novel tracklet similarity measure is formulated based on line geometry. For analysis, we apply non-parametric clustering on the extracted tracklets. Particularly, we adapt the Distance Dependent Chinese Restaurant Process (DD-CRP) to leverage the computed similarities between pairs of tracklets, which ensures the spatial coherence among tracklets in the same cluster. By analyzing the clustering results, we can identify semantic regions in the scene, particularly, the common pathways and their sources/sinks, without any prior information about the scene layout. Qualitative and quantitative experimental evaluation on multiple crowded scenes datasets, principally, the challenging New York Grand Central Station video, demonstrate the state of the art performance of our method.
【Keywords】:
【Paper Link】 【Pages】:3396-3402
【Authors】: Long Lan ; Dacheng Tao ; Chen Gong ; Naiyang Guan ; Zhigang Luo
【Abstract】: Online multi-object tracking (MOT) is challenging: frame-by-frame matching of detection hypotheses to the correct trackers can be difficult. The Hungarian algorithm is the most commonly used online MOT data association method due to its rapid assignment; however, the Hungarian algorithm simply considers associations based on an affinity model. For crowded scenarios, frequently occurring interactions between objects complicate associations, and affinity-based methods usually fail in these scenarios. Here we introduce quadratic pseudo-Boolean optimization (QPBO) to an online MOT model to analyze frequent interactions. Specifically, we formulate two useful interaction types as pairwise potentials in QPBO, a design that benefits our model by exploiting informative interactions and allowing our online tracker to handle complex scenes. The auxiliary interactions result in a non-submodular QPBO, so we accelerate our online tracker by solving the model with a graph cut combined with a simple heuristic method. This combination achieves a reasonable local optimum and, importantly, implements the tracker efficiently. Extensive experiments on publicly available datasets from both static and moving cameras demonstrate the superiority of our method.
【Keywords】:
【Paper Link】 【Pages】:3403-3410
【Authors】: Xiangyuan Lan ; Shengping Zhang ; Pong C. Yuen
【Abstract】: Because of the complementarity of multiple visual cues (features) in appearance modeling, many tracking algorithms attempt to fuse multiple features to improve the tracking performance from two aspects: increasing the representation accuracy against appearance variations and enhancing the discriminability between the tracked target and its background. Since both these two aspects simultaneously contribute to the success of a visual tracker, how to fully unleash the capabilities of multiple features from these two aspects in appearance modeling is a key issue for feature fusion-based visual tracking. To address this problem, different from other feature fusion-based trackers which consider one of these two aspects only, this paper proposes an unified feature learning framework which simultaneously exploits both the representation capability and the discriminability of multiple features for visual tracking. In particular, the proposed feature learning framework is capable of: 1) learning robust features by separating out corrupted features for accurate feature representation, 2) seamlessly imposing the discriminabiltiy of multiple visual cues into feature learning, and 3) fusing features by exploiting their shared and feature-specific discriminative information. Extensive experiment results on challenging videos show that the proposed tracker performs favourably against other ten state-of-the-art trackers.
【Keywords】:
【Paper Link】 【Pages】:3411-3417
【Authors】: Xin Li ; Fan Yang ; Leiting Chen ; Hongbin Cai
【Abstract】: Over the past decades, numerous theories and studies have demonstrated that salient objects in different scenes often share some properties in common that make them visually stand out from their surroundings, and thus can be processed in finer details. In this paper, we propose a novel method for salient object detection that involves the transfer of the annotations from an existing example onto an input image. Our method, which is based on the low-level saliency features of each pixel, estimates dense pixel-wise correspondences between the input image and an example image, and then integrates high-level concepts to produce an initial saliency map. Finally, a coarse-to-fine optimization framework is proposed to generate uniformly highlighted salient objects. Qualitatively and quantitatively experiments on six popular benchmark datasets validate that our approach greatly outperforms the state-of-the-art algorithms and recently published works.
【Keywords】:
【Paper Link】 【Pages】:3418-3424
【Authors】: Wei Liang ; Yibiao Zhao ; Yixin Zhu ; Song-Chun Zhu
【Abstract】: In this paper, we present a probabilistic approach to explicitly infer containment relations between objects in 3D scenes. Given an input RGB-D video, our algorithm quantizes the perceptual space of a 3D scene by reasoning about containment relations over time. At each frame, we represent the containment relations in space by a containment graph, where each vertex represents an object and each edge represents a containment relation. We assume that human actions are the only cause that leads to containment relation changes over time, and classify human actions into four types of events: move-in, move-out, no change and paranormal-change. Here, paranormal-change refers to the events that are physically infeasible, and thus are ruled out through reasoning. A dynamic programming algorithm is adopted to finding both the optimal sequence of containment relations across the video, and the containment relation changes between adjacent frames. We evaluate the proposed method on our dataset with 1326 video clips taken in 9 indoor scenes, including some challenging cases, such as heavy occlusions and diverse changes of containment relations. The experimental results demonstrate good performance on the dataset.
【Keywords】:
【Paper Link】 【Pages】:3425-3431
【Authors】: Xiaobai Liu ; Yadong Mu ; Liang Lin
【Abstract】: This paper presents a stochastic grammar for fine-grained 3D scene reconstruction from a single image. At the heart of our approach is a small number of grammar rules that can describe the most common geometric structures, e.g., two straights lines being co-linear or orthogonal, or that a line lying on a planar region etc. With these grammar rules, we re-frame single-view 3D reconstruction problem as jointly solving two coupled sub-tasks: i) segmenting of image entities, e.g. planar regions, straight edge segments, and ii) optimizing pixel-wise 3D scene model through the application of grammar rules over image entities. To reconstruct a new image, we design an efficient hybrid Monte Carlo (HMC) algorithm to simulate Markov Chain walking towards a posterior distribution. Our algorithm utilizes two iterative dynamics: i) Hamiltonian Dynamics that makes proposals along the gradient direction to search the continuous pixel-wise 3D scene model; and ii) Cluster Dynamics, that flip the colors of clusters of pixels to form planar region partition. Following the Metropolis-hasting principle, these dynamics not only make distant proposals but also guarantee detail-balance and fast convergence. Results with comparisons on public image dataset show that our method clearly outperforms the alternate state-of-the-art single-view reconstruction methods.
【Keywords】:
【Paper Link】 【Pages】:3432-3438
【Authors】: Yao Lu
【Abstract】: The outputs of a trained neural network contain much richer information than just a one-hot classifier. For example, a neural network might give an image of a dog the probability of one in a million of being a cat but it is still much larger than the probability of being a car. To reveal the hidden structure in them, we apply two unsupervised learning algorithms, PCA and ICA, to the outputs of a deep Convolutional Neural Network trained on the ImageNet of 1000 classes. The PCA/ICA embedding of the object classes reveals their visual similarity and the PCA/ICA components can be interpreted as common visual features shared by similar object classes. For an application, we proposed a new zero-shot learning method, in which the visual features learned by PCA/ICA are employed. Our zero-shot learning method achieves the state-of-the-art results on the ImageNet of over 20000 classes.
【Keywords】:
【Paper Link】 【Pages】:3439-3445
【Authors】: Zhanglin Peng ; Ruimao Zhang ; Xiaodan Liang ; Xiaobai Liu ; Liang Lin
【Abstract】: This paper addresses the problem of geometric scene parsing, i.e. simultaneously labeling geometric surfaces (e.g. sky, ground and vertical plane) and determining the interaction relations (e.g. layering, supporting, siding and affinity) between main regions. This problem is more challenging than the traditional semantic scene labeling, as recovering geometric structures necessarily requires the rich and diverse contextual information. To achieve these goals, we propose a novel recurrent neural network model, named Hierarchical Long Short-Term Memory (H-LSTM). It contains two coupled sub-networks: the Pixel LSTM (P-LSTM) and the Multi-scale Super-pixel LSTM (MS-LSTM) for handling the surface labeling and relation prediction, respectively. The two sub-networks provide complementary information to each other to exploit hierarchical scene contexts, and they are jointly optimized for boosting the performance. Our extensive experiments show that our model is capable of parsing scene geometric structures and outperforming several state-of-the-art methods by large margins. In addition, we show promising 3D reconstruction results from the still images based on the geometric parsing.
【Keywords】:
【Paper Link】 【Pages】:3446-3453
【Authors】: Babak Saleh ; Ahmed M. Elgammal ; Jacob Feldman
【Abstract】: Deep artificial neural networks have made remarkable progress in different tasks in the field of computer vision. However, the empirical analysis of these models and investigation of their failure cases has received attention recently. In this work, we show that deep learning models cannot generalize to atypical images that are substantially different from training images. This is in contrast to the superior generalization ability of the visual system in the human brain. We focus on Convolutional Neural Networks (CNN) as the state-of-the-art models in object recognition and classification; investigate this problem in more detail, and hypothesize that training CNN models suffer from unstructured loss minimization. We propose computational models to improve the generalization capacity of CNNs by considering how typical a training image looks like. By conducting an extensive set of experiments we show that involving a typicality measure can improve the classification results on a new set of images by a large margin. More importantly, this significant improvement is achieved without fine-tuning the CNN model on the target image set.
【Keywords】:
【Paper Link】 【Pages】:3454-3461
【Authors】: Tianmin Shu ; Michael S. Ryoo ; Song-Chun Zhu
【Abstract】: In this paper, we present an approach for robot learning of social affordance from human activity videos. We consider the problem in the context of human-robot interaction: our approach learns structural representations of human-human (and human-object-human) interactions, describing how body-parts of each agent move with respect to each other and what spatial relations they should maintain to complete each sub-event (i.e., sub-goal). This enables the robot to infer its own movement in reaction to the human body motion, allowing it to naturally replicate such interactions.
【Keywords】:
【Paper Link】 【Pages】:3462-3468
【Authors】: Jivko Sinapov ; Priyanka Khante ; Maxwell Svetlik ; Peter Stone
【Abstract】: This paper proposes a novel framework that enables a robot to learn ordinal object relations. While most related work focuses on classifying objects into discrete categories, such approaches cannot learn object properties (e.g., weight, height, size, etc.) that are context-specific and relative to other objects. To address this problem, we propose that a robot should learn to order objects based on ordinal object relations. In our experiments, the robot explored a set of 32 objects that can be ordered by three properties: height, weight, and width. Next, the robot used unsupervised learning to discover multiple ways that the objects can be ordered based on the haptic and proprioceptive perceptions detected while exploring the objects. Following, the robot's model was presented with labeled object series, allowing it to ground the three ordinal relations in terms of how similar they are to the orders discovered during the unsupervised stage. Finally, the grounded models were used to recognize whether new object series were ordered by any of the three properties as well as to correctly insert additional objects into an existing series.
【Keywords】:
【Paper Link】 【Pages】:3469-3476
【Authors】: Hang Su ; Yinpeng Dong ; Jun Zhu ; Haibin Ling ; Bo Zhang
【Abstract】: Exploring crowd dynamics is essential in understanding crowd scenes, which still remains as a challenging task due to the nonlinear characteristics and coherent spatio-temporal motion patterns in crowd behaviors. To address these issues, we present a Coherent Long Short Term Memory (cLSTM) network to capture the nonlinear crowd dynamics by learning an informative representation of crowd motions, which facilitates the critical tasks in crowd scene analysis. By describing the crowd motion patterns with a cloud of keypoint tracklets, we explore the nonlinear crowd dynamics embedded in the tracklets with a stacked LSTM model, which is further improved to capture the collective properties by introducing a coherent regularization term; and finally, we adopt an unsupervised encoder-decoder framework to learn a hidden feature for each input tracklet that embeds its inherent dynamics. With the learnt features properly harnessed, crowd scene understanding is conducted effectively in predicting the future paths of agents, estimating group states, and classifying crowd events. Extensive experiments on hundreds of public crowd videos demonstrate that our method is state-of-the-art performance by exploring the coherent spatio-temporal structures in crowd behaviors.
【Keywords】:
【Paper Link】 【Pages】:3477-3483
【Authors】: Jesse Thomason ; Jivko Sinapov ; Maxwell Svetlik ; Peter Stone ; Raymond J. Mooney
【Abstract】: Grounded language learning bridges words like red and square with robot perception. The vast majority of existing work in this space limits robot perception to vision. In this paper, we build perceptual models that use haptic, auditory, and proprioceptive data acquired through robot exploratory behaviors to go beyond vision. Our system learns to ground natural language words describing objects using supervision from an interactive human-robot I Spy game. In this game, the human and robot take turns describing one object among several, then trying to guess which object the other has described. All supervision labels were gathered from human participants physically present to play this game with a robot. We demonstrate that our multi-modal system for grounding natural language outperforms a traditional, vision-only grounding framework by comparing the two on the "I Spy" task. We also provide a qualitative analysis of the groundings learned in the game, visualizing what words are understood better with multi-modal sensory information as well as identifying learned word meanings that correlate with physical object properties (e.g. "small" negatively correlates with object weight).
【Keywords】:
【Paper Link】 【Pages】:3484-3490
【Authors】: Jingwen Wang ; Jianlong Fu ; Yong Xu ; Tao Mei
【Abstract】: Visual sentiment analysis aims to automatically recognize positive and negative emotions from images. There are three main challenges, including large intra-class variance, fine-grained image categories, and scalability. Most existing methods predominantly focus on one or two challenges, which has limited their performance. In this paper, we propose a novel visual sentiment analysis approach with deep coupled adjective and noun neural networks. Specifically, to reduce the large intra-class variance, we first learn a shared middle-level sentiment representation by jointly learning an adjective and a noun deep neural network with weak label supervision. Second, based on the learned sentiment representation, a prediction network is further optimized to deal with the subtle differences which often exist in the fine-grained image categories. The three networks are trained in end-to-end manner, where the middle-level representation learned in previous two networks can guide the sentiment network to achieve high performance and fast convergence. Third, we generalize the training with mutual supervision between the learned adjective and noun networks by a Rectified Kullback-Leibler loss (ReKL), when the adjective and noun labels are not available. Extensive experiments on two widely-used datasets show that our method outperforms the state-of-the-art on SentiBank dataset with 10.2% accuracy gain and surpasses the previous best approach on Twitter dataset with clear margin.
【Keywords】:
【Paper Link】 【Pages】:3491-3497
【Authors】: Shu Wang ; Shaoting Zhang ; Wei Liu ; Dimitris N. Metaxas
【Abstract】: In this paper, we propose a novel visual tracking framework that intelligently discovers reliable patterns from a wide range of video to resist drift error for long-term tracking tasks. First, we design a Discrete Fourier Transform (DFT) based tracker which is able to exploit a large number of tracked samples while still ensures real-time performance. Second, we propose a clustering method with temporal constraints to explore and memorize consistent patterns from previous frames, named as reliable memories. By virtue of this method, our tracker can utilize uncontaminated information to alleviate drifting issues. Experimental results show that our tracker performs favorably against other state-of-the-art methods on benchmark datasets. Furthermore, it is significantly competent in handling drifts and able to robustly track challenging long videos over 4,000 frames, while most of others lose track at early frames.
【Keywords】:
【Paper Link】 【Pages】:3498-3504
【Authors】: Xiaoyang Wang ; Qiang Ji
【Abstract】: Attribute based object recognition performs object recognition using the semantic properties of the object. Unlike the existing approaches that treat attributes as a middle level representation and require to estimate the attributes during testing, we propose to incorporate the hidden attributes, which are the attributes used only during training to improve model learning and are not needed during testing. To achieve this goal, we develop two different approaches to incorporate hidden attributes. The first approach utilizes hidden attributes as additional information to improve the object classification model. The second approach further exploits the semantic relationships between the objects and the hidden attributes. Experiments on benchmark data sets demonstrate that both approaches can effectively improve the learning of the object classifiers over the baseline models that do not use attributes, and their combination reaches the best performance. Experiments also show that the proposed approaches outperform both state of the art methods that use attributes as middle level representation and the approaches that learn the classifiers with hidden information.
【Keywords】:
【Paper Link】 【Pages】:3505-3512
【Authors】: Inbal Wiesel-Kapah ; Gal A. Kaminka ; Guy Hachmon ; Noa Agmon ; Ido Bachelet
【Abstract】: Molecular robots (nanobots) are being developed for biomedical applications, e.g., to deliver medications without worrying about side-effects. Future treatments will require swarms of heterogeneous nanobots. We present a novel approach to generating such swarms from a treatment program. A compiler translates medications, written in a rule-based language, into specifications of a swarm built by specializing generic nanobot platforms to specific pay-loads and action-triggering behavior. The mixture of nanobots, when deployed, carries out the treatment program. We describe the medication programming language, and the associated compiler. We prove the validity of the compiler output, and report on in-vitro experiments using generated nanobot swarms.
【Keywords】:
【Paper Link】 【Pages】:3513-3521
【Authors】: Lawson L. S. Wong ; Thanard Kurutach ; Tomás Lozano-Pérez ; Leslie Pack Kaelbling
【Abstract】: To accomplish tasks in human-centric indoor environments, agents need to represent and understand the world in terms of objects and their attributes. We consider how to acquire such a world model via noisy perception and maintain it over time, as objects are added, changed, and removed in the world. Previous work framed this as multiple-target tracking problem, where objects are potentially in motion at all times. Although this approach is general, it is computationally expensive. We argue that such generality is not needed in typical world modeling tasks, where objects only change state occasionally. More efficient approaches are enabled by restricting ourselves to such semi-static environments. We consider a previously-proposed clustering-based world modeling approach that assumed static environments, and extend it to semi-static domains by applying a dependent Dirichlet process (DDP) mixture model. We derive a novel MAP inference algorithm under this model, subject to data association constraints. We demonstrate our approach improves computational performance for world modeling in semi-static environments.
【Keywords】:
【Paper Link】 【Pages】:3522-3529
【Authors】: Xue Yang ; Fei Han ; Hua Wang ; Hao Zhang
【Abstract】: Sparse representation has been widely studied in visual tracking, which has shown promising tracking performance. Despite a lot of progress, the visual tracking problem is still a challenging task due to appearance variations over time. In this paper, we propose a novel sparse tracking algorithm that well addresses temporal appearance changes, by enforcing template representability and temporal consistency (TRAC). By modeling temporal consistency, our algorithm addresses the issue of drifting away from a tracking target. By exploring the templates' long-term-short-term representability, the proposed method adaptively updates the dictionary using the most descriptive templates, which significantly improves the robustness to target appearance changes. We compare our TRAC algorithm against the state-of-the-art approaches on 12 challenging benchmark image sequences. Both qualitative and quantitative results demonstrate that our algorithm significantly outperforms previous state-of-the-art trackers.
【Keywords】:
【Paper Link】 【Pages】:3530-3537
【Authors】: Hang Yin ; Patricia Alves-Oliveira ; Francisco S. Melo ; Aude Billard ; Ana Paiva
【Abstract】: This paper contributes a novel framework that enables a robotic agent to efficiently learn and synthesize believable handwriting motion. We situate the framework as a foundation with the goal of allowing children to observe, correct and engage with the robot to learn themselves the handwriting skill. The framework adapts the principle behind ensemble methods - where improved performance is obtained by combining the output of multiple simple algorithms - in an inverse optimal control problem. This integration addresses the challenges of rapid extraction and representation of multiple-mode motion trajectories, with the cost forms which are transferable and interpretable in the development of the robot compliance control. It also introduces the incorporation of a human movement inspired feature, which provides intuitive motion modulation to generalize the synthesis with poor robotic written samples for children to identify and correct. We present the results on the success of synthesizing a variety of natural-looking motion samples based upon the learned cost functions. The framework is validated by a user study, where the synthesized dynamical motion is shown to be hard to distinguish from the real human handwriting.
【Keywords】:
【Paper Link】 【Pages】:3538-3544
【Authors】: Dingwen Zhang ; Deyu Meng ; Long Zhao ; Junwei Han
【Abstract】: Weakly-supervised object detection (WOD) is a challenging problems in computer vision. The key problem is to simultaneously infer the exact object locations in the training images and train the object detectors, given only the training images with weak image-level labels. Intuitively, by simulating the selective attention mechanism of human visual system, saliency detection technique can select attractive objects in scenes and thus is a potential way to provide useful priors for WOD. However, the way to adopt saliency detection in WOD is not trivial since the detected saliency region might be possibly highly ambiguous in complex cases. To this end, this paper first comprehensively analyzes the challenges in applying saliency detection to WOD. Then, we make one of the earliest efforts to bridge saliency detection to WOD via the self-paced curriculum learning, which can guide the learning procedure to gradually achieve faithful knowledge of multi-class objects from easy to hard. The experimental results demonstrate that the proposed approach can successfully bridge saliency detection and WOD tasks and achieve the state-of-the-art object detection results under the weak supervision.
【Keywords】:
【Paper Link】 【Pages】:3545-3551
【Authors】: Yaqing Zhang ; Xi Li ; Liming Zhao ; Zhongfei Zhang
【Abstract】: In this paper, we propose an end-to-end deep correspondence structure learning (DCSL) approach to address the cross-camera person-matching problem in the person re-identification task. The proposed DCSL approach captures the intrinsic structural information on persons by learning a semantics-aware image representation based on convolutional neural networks, which adaptively learns discriminative features for person identification. Furthermore, the proposed DCSL approach seeks to adaptively learn a hierarchical data-driven feature matching function which outputs the matching correspondence results between the learned semantics-aware image representations for a person pair. Finally, we set up a unified end-to-end deep learning scheme to jointly optimize the processes of semantics-aware image representation learning and cross-person correspondence structure learning, leading to more reliable and robust person re-identification results in complicated scenarios. Experimental results on several benchmark datasets demonstrate the effectiveness of our approach against the state-of-the-art approaches.
【Keywords】:
【Paper Link】 【Pages】:3552-3559
【Authors】: Xiaoke Zhu ; Xiao-Yuan Jing ; Fei Wu ; Hui Feng
【Abstract】: Video-based person re-identification (re-id) is an important application in practice. However, only a few methods have been presented for this problem. Since large variations exist between different pedestrian videos, as well as within each video, it's challenging to conduct re-identification between pedestrian videos. In this paper, we propose a simultaneous intra-video and inter-video distance learning (SI2DL) approach for video-based person re-id. Specifically, SI2DL simultaneously learns an intra-video distance metric and an inter-video distance metric from the training videos. The intra-video distance metric is to make each video more compact, and the inter-video one is to make that the distance between two truly matching videos is smaller than that between two wrong matching videos. To enhance the discriminability of learned metrics, we design a video relationship model, i.e., video triplet, for SI2DL. Experiments on the public iLIDS-VID and PRID 2011 image sequence datasets show that our approach achieves the state-of-the-art performance.
【Keywords】:
【Paper Link】 【Pages】:3560-3568
【Authors】: Ankit Anand ; Aditya Grover ; Mausam ; Parag Singla
【Abstract】: An important approach for efficient inference in probabilistic graphical models exploits symmetries among objects in the domain. Symmetric variables (states) are collapsed into meta-variables (metastates) and inference algorithms are run over the lifted graphical model instead of the flat one. Our paper extends existing definitions of symmetry by introducing the novel notion of contextual symmetry. Two states that are not globally symmetric, can be contextually symmetric under some specific assignment to a subset of variables, referred to as the context variables. Contextual symmetry subsumes previous symmetry definitions and can represent a large class of symmetries not representable earlier. We show how to compute contextual symmetries by reducing it to the problem of graph isomorphism. We extend previous work on exploiting symmetries in the MCMC framework to the case of contextual symmetries. Our experiments on several domains of interest demonstrate that exploiting contextual symmetries can result in significant computational gains.
【Keywords】:
【Paper Link】 【Pages】:3569-3576
【Authors】: Supratik Chakraborty ; Kuldeep S. Meel ; Moshe Y. Vardi
【Abstract】: Probabilistic inference via model counting has emerged as a scalable technique with strong formal guarantees, thanks to recent advances in hashing-based approximate counting. State-of-the-art hashing-based counting algorithms use an NP oracle, such that the number of oracle invocations grows linearly in the number of variables n in the input constraint. We present a new approach to hashing-based approximate model counting in which the number of oracle invocations grows logarithmically in n, while still providing strong theoretical guarantees. We use this technique to design an algorithm for #CNF with strongly probably approximately correct (SPAC) guarantees, i.e. PAC guarantee plus expected return value matching the exact model count. Our experiments show that this algorithm outperforms state-of-the-art techniques for approximate counting by 1-2 orders of magnitude in running time. We also show that our algorithm can be easily adapted to give a new fully polynomial randomized approximation scheme (FPRAS) for #DNF.
【Keywords】:
【Paper Link】 【Pages】:3577-3583
【Authors】: Bryant Chen ; Judea Pearl ; Elias Bareinboim
【Abstract】: In this paper, we extend graph-based identification methods by allowing background knowledge in the form of non-zero parameter values. Such information could be obtained, for example, from a previously conducted randomized experiment, from substantive understanding of the domain, or even an identification technique. To incorporate such information systematically, we propose the addition of auxiliary variables to the model, which are constructed so that certain paths will be conveniently cancelled. This cancellation allows the auxiliary variables to help conventional methods of identification (e.g., single-door criterion, instrumental variables, half-trek criterion), as well as model testing (e.g., d-separation, over-identification). Moreover, by iteratively alternating steps of identification and adding auxiliary variables, we can improve the power of existing identification methods via a bootstrapping approach that does not require external knowledge. We operationalize this method for simple instrumental sets (a generalization of instrumental variables) and show that the resulting method is able to identify at least as many models as the most general identification method for linear systems known to date. We further discuss the application of auxiliary variables to the tasks of model testing and z-identification.
【Keywords】:
【Paper Link】 【Pages】:3584-3590
【Authors】: Cong Chen ; Changhe Yuan ; Chao Chen
【Abstract】: M-Modes for graphical models is the problem of finding top M label configurations of highest probability in their local neighborhoods. The state-of-the-art method for solving M-Modes is a dynamic programming algorithm which computes global modes by first computing local modes of each subgraph and then search through all their consistent combinations. A drawback of the algorithm is that most of its time is wasted on computing local modes that are never used in global modes. This paper introduces new algorithms that directly search the space of consistent local modes in finding the global modes, which is enabled by a novel search operator designed to search a subgraph of variables at each time. As a result, the search algorithms only need to generate and verify a small number of local modes and can hence lead to significant improvement in efficiency and scalability.
【Keywords】:
【Paper Link】 【Pages】:3591-3599
【Authors】: Rodrigo de Salvo Braz ; Ciaran O'Reilly ; Vibhav Gogate ; Rina Dechter
【Abstract】: We present SGDPLL(T), an algorithm that solves (among many other problems) probabilistic inference modulo theories, that is, inference problems over probabilistic models defined via a logic theory provided as a parameter (currently, propositional, equalities on discrete sorts, and inequalities, more specifically difference arithmetic, on bounded integers). While many solutions to probabilistic inference over logic representations have been proposed, SGDPLL(T) is simultaneously (1) lifted, (2) exact and (3) modulo theories, that is, parameterized by a background logic theory. This offers a foundation for extending it to rich logic languages such as data structures and relational data. By lifted, we mean algorithms with constant complexity in the domain size (the number of values that variables can take). We also detail a solver for summations with difference arithmetic and show experimental results from a scenario in which SGDPLL(T) is much faster than a state-of-the-art probabilistic solver.
【Keywords】:
【Paper Link】 【Pages】:3600-3608
【Authors】: Daisuke Hatano ; Takuro Fukunaga ; Ken-ichi Kawarabayashi
【Abstract】: The budget allocation problem is an optimization problem arising from advertising planning. In the problem, an advertiser has limited budgets to allocate across media, and seeks to optimize the allocation such that the largest fraction of customers can be influenced. It is known that this problem admits a (1 – 1/e)-approximation algorithm. However, no previous studies on this problem considered adjusting the allocation adaptively based upon the effect of the past campaigns, which is a usual strategy in the real setting. Our main contribution in this paper is to analyze adaptive strategies for the budget allocation problem. We define a greedy strategy, referred to as the insensitive policy, and then give a provable performance guarantee. This result is obtained by extending the adaptive submodularity, which is a concept studied in the context of active learning and stochastic optimization, to the functions over an integer lattice.
【Keywords】:
【Paper Link】 【Pages】:3609-3615
【Authors】: Shamin Kinathil ; Scott Sanner ; Sanmay Das ; Nicolás Della Penna
【Abstract】: Market-makers serve an important role as providers of liquidity and order in financial markets, particularly during periods of high volatility. Optimal market-makers solve a sequential decision making problem, where they face an exploration versus exploitation dilemma at each time step. A belief state MDP based solution was presented by Das and Magdon-Ismail [NIPS, 2008]. This solution however, was closely tied to the choice of a Gaussian belief state prior and did not take asset inventory into consideration when calculating an optimal policy. In this work we introduce a novel continuous state POMDP framework which is the first to solve, exactly and in closed-form, the optimal market making problem with inventory, fixed asset value, arbitrary belief state priors, trader models and reward functions via symbolic dynamic programming. We use this novel model and solution to show that sequentially optimal policies are heavily inventory-dependent and calculate policies that operate with bounded loss guarantees under a variety of market models and conditions.
【Keywords】:
【Paper Link】 【Pages】:3616-3622
【Authors】: Steffen Michels ; Arjen Hommersom ; Peter J. F. Lucas
【Abstract】: Probabilistic logics, especially those based on logic programming (LP), are gaining popularity as modelling and reasoning tools, since they combine the power of logic to represent knowledge with the ability of probability theory to deal with uncertainty. In this paper, we propose a hybrid extension for probabilistic logic programming, which allows for exact inference for a much wider class of continuous distributions than existing extensions. At the same time, our extension allows one to compute approximations with bounded and arbitrarily small error. We propose a novel anytime algorithm exploiting the logical and continuous structure of distributions and experimentally show that our algorithm is, for typical relational problems, competitive with state-of-the-art sampling algorithms and outperforms them by far if rare events with deterministic structure are provided as evidence, despite the fact that it provides much stronger guarantees.
【Keywords】:
【Paper Link】 【Pages】:3623-3629
【Authors】: Nico Potyka ; Erman Acar ; Matthias Thimm ; Heiner Stuckenschmidt
【Abstract】: We propose a probabilistic-logical framework for group decision-making. Its main characteristic is that we derive group preferences from agents' beliefs and utilities rather than from their individual preferences as done in social choice approaches. This can be more appropriate when the individual preferences hide too much of the individuals' opinions that determined their preferences. We introduce three preference relations and investigate the relationships between the group preferences and individual and subgroup preferences.
【Keywords】:
【Paper Link】 【Pages】:3630-3636
【Authors】: Min Wen ; Ufuk Topcu
【Abstract】: We consider a controller synthesis problem in turn-based stochastic games with both a qualitative linear temporal logic (LTL) constraint and a quantitative discounted-sum objective. For each case in which the LTL specification is realizable and can be equivalently transformed into a deterministic Buchi automaton, we show that there always exists a memoryless almost-sure winning strategy that is epsilon-optimal with respect to the discounted-sum objective for any arbitrary positive epsilon. Building on the idea of the R-MAX algorithm, we propose a probably approximately correct (PAC) learning algorithm that can learn such a strategy efficiently in an online manner with a-priori unknown reward functions and unknown transition distributions. To the best of our knowledge, this is the first result on PAC learning in stochastic games with independent quantitative and qualitative objectives.
【Keywords】:
【Paper Link】 【Pages】:3637-3645
【Authors】: Yi Wu ; Lei Li ; Stuart Russell ; Rastislav Bodík
【Abstract】: A probabilistic program defines a probability measure over its semantic structures. One common goal of probabilistic programming languages (PPLs) is to compute posterior probabilities for arbitrary models and queries, given observed evidence, using a generic inference engine. Most PPL inference engines — even the compiled ones — incur significant runtime interpretation overhead, especially for contingent and open-universe models. This paper describes Swift, a compiler for the BLOG PPL. Swift-generated code incorporates optimizations that eliminate interpretation overhead, maintain dynamic dependencies efficiently, and handle memory management for possible worlds of varying sizes. Experiments comparing Swift with other PPL engines on avariety of inference problems demonstrate speedups ranging from 12x to326x.
【Keywords】:
【Paper Link】 【Pages】:3646-3653
【Authors】: Li Zhou ; Emma Brunskill
【Abstract】: Personalized recommendations for new users, also known as the cold-start problem, can be formulated as a contextual bandit problem. Existing contextual bandit algorithms generally rely on features alone to capture user variability. Such methods are inefficient in learning new users' interests. In this paper we propose Latent Contextual Bandits. We consider both the benefit of leveraging a set of learned latent user classes for new users, and how we can learn such latent classes from prior users. We show that our approach achieves a better regret bound than existing algorithms. We also demonstrate the benefit of our approach using a large real world dataset and a preliminary user study.
【Keywords】:
【Paper Link】 【Pages】:3654-3661
【Authors】: Zheng Zhou ; Kan Li ; Xiangjian He ; Mengmeng Li
【Abstract】: Recognizing multiple mixed group activities from one still image is not a hard problem for humans but remains highly challenging for computer recognition systems. When modeling interactions among multiple units (i.e., more than two groups or persons), the existing approaches tend to divide them into interactions between pairwise units. However, no mathematical evidence supports this transformation. Therefore, these approaches' performance is limited on images containing multiple activities. In this paper, we propose a generative model to provide a more reasonable interpretation for the mixed group activities contained in one image. We design a four level structure and convert the original intra-level interactions into inter-level interactions, in order to implement both interactions among multiple groups and interactions among multiple persons within a group. The proposed four-level structure makes our model more robust against the occlusion and overlap of the visible poses in images. Experimental results demonstrate that our model makes good interpretations for mixed group activities and outperforms the state-of-the-art methods on the Collective Activity Classification dataset.
【Keywords】:
【Paper Link】 【Pages】:3662-3669
【Authors】: Garrett Andersen ; Vincent Conitzer
【Abstract】: In highly anonymous environments such as the Internet, many applications suffer from the fact that a single user can pose as multiple users. Indeed, presumably many potential applications do not even get off the ground as a result. Consider the example of an online vote. Requiring voters to provide identifying information, to the extent that this is even feasible, can significantly deter participation. On the other hand, not doing so makes it possible for a single individual to vote more than once, so that the result may become almost meaningless. (A quick web search will reveal many examples of Internet polls with bizarre outcomes.) CAPTCHAs may prevent running a program that votes many times, but they do nothing to prevent a single user from voting many times by hand. In this paper, we propose ATUCAPTS (Automated Tests That a User Cannot Pass Twice Simultaneously) as a solution. ATUCAPTS are automatically generated tests such that it is (1) easy for a user to pass one instance, but (2) extremely difficult for a user to pass two instances at the same time. Thus, if it is feasible to require all users to take such a test at the same time, we can verify that no user holds more than one account. We propose a specific class of ATUCAPTS and present the results of a human subjects study to validate that they satisfy the two properties above. We also introduce several theoretical models of how well an attacker might perform and show that these models still allow for good performance on both (1) and (2) with reasonable test lengths.
【Keywords】:
【Paper Link】 【Pages】:3670-3676
【Authors】: Ehsan Mohammady Ardehaly ; Aron Culotta
【Abstract】: Learning from Label Proportions (LLP) is a machine learning problem in which the training data consist of bags of instances, and only the class label distribution for each bag is known. In some domains label proportions are readily available; for example, by grouping social media users by location, one can use census statistics to build a classifier for user demographics. However, label proportions are unavailable in many domains, such as product review sites. The goal of this paper is to determine whether an LLP classifier fit in one domain can be modified to classify instances from another domain. To do so, we propose a domain adaptation algorithm that uses an LLP model fit on the source domain to generate label proportions for the target domain. A new LLP model is then fit on the target domain, and this self-training process is repeated to adapt the model from source to target. Our experiments on five diverse tasks indicate an 11% average absolute improvement in accuracy as compared to using LLP without domain adaptation. In contrast to existing domain adaptation algorithms, our approach requires only label proportions in the source domain, and the results suggest that the approach is effective even when the target domain is substantially different from the source domain.
【Keywords】:
【Paper Link】 【Pages】:3677-3683
【Authors】: Qing Bao ; William K. Cheung ; Jiming Liu
【Abstract】: Existing diffusion models for social networks often assume that the activation of a node depends independently on their parents' activations. Some recent work showed that incorporating the structural and behavioral dependency among the parent nodes allows more accurate diffusion models to be inferred. In this paper, we postulate that the latent temporal activation patterns (or motifs) of nodes of different social roles form the underlying information diffusion mechanisms generating the information cascades observed over a social network. We formulate the inference of the temporal activation motifs and a corresponding motif-based diffusion model under a unified probabilistic framework. A two-level EM algorithm is derived so as to infer the diffusion-specific motifs and the diffusion probabilities simultaneously. We applied the proposed model to several real-world datasets with significant improvement on modelling accuracy. We also illustrate how the inferred motifs can be interpreted as the underlying mechanisms causing the diffusion process to happen in different social networks.
【Keywords】:
【Paper Link】 【Pages】:3684-3690
【Authors】: Salem Benferhat ; Zied Bouraoui ; Madalina Croitoru ; Odile Papini ; Karim Tabia
【Abstract】: Repair based techniques are a standard way of dealing with inconsistency in the context of ontology based data access. We propose a novel non-objection inference relation (along with its variants) where a query is considered as valid if it follows from at least one repair and it is consistent with all the repairs. These inferences are strictly more productive than universal inference while preserving the consistency of its set of conclusions. We study the productivity and properties of the new inferences. We also give experimental results of the proposed non-objection inference.
【Keywords】:
【Paper Link】 【Pages】:3691-3697
【Authors】: Himanshu S. Bhatt ; Arun Rajkumar ; Shourya Roy
【Abstract】: Owing to the tremendous increase in the volume and variety of user generated content, train-once-apply-forever models are insufficient for supervised learning tasks. Thus, developing algorithms that adapt across domains by leveraging data from multiple domains is critical. However, existing adaptation algorithms often fail to identify the right sources to use for adaptation. In this work, we present a novel multi-source iterative domain adaptation algorithm (MSIDA) that leverages knowledge from selective sources to improve the performance in a target domain. The algorithm first chooses the best K sources from possibly numerous existing domains taking into account both similarity and complementarity properties of the domains. Then it learns target specific features in an iterative manner building on the common shared representations from the source domains. We give theoretical justifications for our source selection procedure and also give mistake bounds for the MSIDA algorithm. Experimental results justify the theory as MSIDA significantly outperforms existing cross-domain classification approaches on the real world and benchmark datasets.
【Keywords】:
【Paper Link】 【Pages】:3698-3704
【Authors】: Yi Chang ; Jiliang Tang ; Dawei Yin ; Makoto Yamada ; Yan Liu
【Abstract】: The popularity of social media shatters the barrier for online users to create and share information at any place at any time. As a consequence, it has become increasing difficult to locate relevance information about an entity. Timeline has been proven to provide an effective and efficient access to understand an entity by displaying a list of episodes about the entity in chronological order. However, summarizing the timeline about an entity with social media data faces new challenges. First, key timeline episodes about the entity are typically unavailable in existing social media services. Second, the short, noisy and informal nature of social media posts determines that only content-based summarization could be insufficient. In this paper, we investigate the problem of timeline summarization and propose a novel framework Timeline-Sumy, which consists of episode detecting and summary ranking. In episode detecting, we explicitly model temporal information with life cycle models to detect timeline episodes since episodes usually exhibit sudden-rise-and-heavy-tail patterns on time-series. In summary ranking, we rank social media posts in each episode via a learning-to-rank approach. The experimental results on social media datasets demonstrate the effectiveness of the proposed framework.
【Keywords】:
【Paper Link】 【Pages】:3705-3711
【Authors】: Gong Cheng ; Cheng Jin ; Yuzhong Qu
【Abstract】: The rapid growth of open data on the Web promotes the development of data portals that facilitate finding useful datasets. To help users quickly inspect a dataset found in a portal, we propose to summarize its contents and generate a hierarchical grouping of entities connected by relations. Our generic approach, called HIEDS, considers coverage of dataset, height of hierarchy, cohesion within groups, overlap between groups, and homogeneity of groups, and integrates these configurable factors into a combinatorial optimization problem to solve. We present an efficient solution, to serve users with dynamically configured summaries with acceptable latency. We systematically experiment with our approach on real-world RDF datasets.
【Keywords】:
【Paper Link】 【Pages】:3712-3718
【Authors】: Yo Ehara ; Yukino Baba ; Masao Utiyama ; Eiichiro Sumita
【Abstract】: Translation ability is known as one of the most difficult language abilities to measure. A typical method of measuring translation ability involves asking translators to translate sentences and to request professional evaluators to grade the translations. It imposes a heavy burden on both translators and evaluators. In this paper, we propose a practical method for assessing translation ability. Our key idea is to incorporate translators' vocabulary knowledge for translation ability assessment. Our method involves just asking translators to tell if they know given words. Using this vocabulary information, we build a probabilistic model to estimate the translators' vocabulary and translation abilities simultaneously. We evaluated our method in a realistic crowdsourcing translation setting in which there is a great need to measure translators' translation ability to select good translators. The results of our experiments show that the proposed method accurately estimates translation ability and selects translators who have sufficient skills in translating a given sentence. We also found that our method significantly reduces the cost of crowdsourcing translation.
【Keywords】:
【Paper Link】 【Pages】:3719-3725
【Authors】: Ziyu Guan ; Long Chen ; Wei Zhao ; Yi Zheng ; Shulong Tan ; Deng Cai
【Abstract】: Sentiment analysis is one of the key challenges for mining online user generated content. In this work, we focus on customer reviews which are an important form of opinionated content. The goal is to identify each sentence's semantic orientation (e.g. positive or negative) of a review. Traditional sentiment classification methods often involve substantial human efforts, e.g. lexicon construction, feature engineering. In recent years, deep learning has emerged as an effective means for solving sentiment classification problems. A neural network intrinsically learns a useful representation automatically without human efforts. However, the success of deep learning highly relies on the availability of large-scale training data. In this paper, we propose a novel deep learning framework for review sentiment classification which employs prevalently available ratings as weak supervision signals. The framework consists of two steps: (1) learn a high level representation (embedding space) which captures the general sentiment distribution of sentences through rating information; (2) add a classification layer on top of the embedding layer and use labeled sentences for supervised fine-tuning. Experiments on review data obtained from Amazon show the efficacy of our method and its superiority over baseline methods.
【Keywords】:
【Paper Link】 【Pages】:3726-3732
【Authors】: Qi Guo ; Chinmay Kulkarni ; Aniket Kittur ; Jeffrey P. Bigham ; Emma Brunskill
【Abstract】: Formative assessments allow learners to quickly identify knowledge gaps. In traditional educational settings, expert instructors can create assessments, but in informal learning environment, it is difficult for novice learners to self assess because they don't know what they don't know. This paper introduces Questimator, an automated system that generates multiple-choice assessment questions for any topic contained within Wikipedia. Given a topic, Questimator traverses the Wikipedia graph to find and rank related topics, and uses article text to form questions, answers and distractor options. In a study with 833 participants from Mechanical Turk, we found that participants' scores on Questimator-generated quizzes correlated well with their scores on existing online quizzes on topics ranging from philosophy to economics. Also Questimator generates questions with comparable discriminatory power as existing online quizzes. Our results suggest Questimator may be useful for assessing learning in topics for which there is not an existing quiz.
【Keywords】:
【Paper Link】 【Pages】:3733-3739
【Authors】: Takanori Hayashi ; Takuya Akiba ; Yuichi Yoshida
【Abstract】: In a connected graph, spanning tree centralities of a vertex and an edge measure how crucial they are for the graph to be connected. In this paper, we propose efficient algorithms for estimating spanning tree centralities with theoretical guarantees on their accuracy. We experimentally demonstrate that our methods are orders of magnitude faster than previous methods. Then, we propose a novel centrality notion of a vertex, called aggregated spanning tree centrality, which also considers the number of connected components obtained by removing the vertex. We also give an efficient algorithm for estimating aggregated spanning tree centrality. Finally, we experimentally show that those spanning tree centralities are useful to identify vulnerable edges and vertices in infrastructure networks.
【Keywords】:
【Paper Link】 【Pages】:3740-3746
【Authors】: Ruining He ; Chunbin Lin ; Jianguo Wang ; Julian McAuley
【Abstract】: Building successful recommender systems requires uncovering the underlying dimensions that describe the properties of items as well as users' preferences toward them. In domains like clothing recommendation, explaining users' preferences requires modeling the visual appearance of the items in question. This makes recommendation especially challenging, due to both the complexity and subtlety of people's visual preferences, as well as the scale and dimensionality of the data and features involved. Ultimately, a successful model should be capable of capturing considerable variance across different categories and styles, while still modeling the commonalities explained by global structures in order to combat the sparsity (e.g. cold-start), variability, and scale of real-world datasets. Here, we address these challenges by building such structures to model the visual dimensions across different product categories. With a novel hierarchical embedding architecture, our method accounts for both high-level (colorfulness, darkness, etc.) and subtle (e.g. casualness) visual characteristics simultaneously.
【Keywords】:
【Paper Link】 【Pages】:3747-3753
【Authors】: Tatsuya Iwanari ; Naoki Yoshinaga ; Nobuhiro Kaji ; Toshiharu Nishina ; Masashi Toyoda ; Masaru Kitsuregawa
【Abstract】: This paper presents a novel task of ordering given concepts (e.g., London, Paris, and Rome) on the basis of common attribute intensity expressed by a given adjective (e.g., safe) and proposes statistical ordering methods that integrate heterogeneous evidence extracted from text on concept ordering. This study is aimed at deriving collective wisdom on concept ordering from social media text. Solving this task is not only interesting from a sociological perspective but also beneficial in the practical sense for those who want to order unfamiliar entities in terms of subjective attributes that are hard to quantify in order to make correct decisions. Experiments on real-world concepts revealed a strong correlation between orderings obtained by our methods and gold-standard orderings.
【Keywords】:
【Paper Link】 【Pages】:3754-3760
【Authors】: Chris Kedzie ; Fernando Diaz ; Kathleen McKeown
【Abstract】: We present a system based on sequential decision making for the online summarization of massive document streams, such as those found on the web. Given an event of interest (e.g. Boston Marathon bombing), our system is able to filter the stream for relevance and produce a series of short text updates describing the event as it unfolds over time. Unlike previous work, our approach is able to jointly model the relevance, comprehensiveness, novelty, and timeliness required by time-sensitive queries. We demonstrate a 28.3% improvement in summary F1 and a 43.8% improvement in time-sensitive F1 metrics.
【Keywords】:
【Paper Link】 【Pages】:3761-3767
【Authors】: Takuya Konishi ; Tomoharu Iwata ; Kohei Hayashi ; Ken-ichi Kawarabayashi
【Abstract】: Identifying soon-to-be-popular items in web services offers important benefits. We attempt to identify users who can find prospective popular items. Such visionary users are called observers. By adding observers to a favorite user list, they act to find popular items in advance. To identify efficient observers, we propose a feature selection based framework. This uses a classifier to predict item popularity, where the input features are a set of users who adopted an item before others. By training the classifier with sparse and non-negative constraints, observers are extracted as users whose parameters take a non-zero value. In experiments, we test our approach using real social bookmark datasets. The results demonstrate that our approach can find popular items in advance more effectively than baseline methods.
【Keywords】:
【Paper Link】 【Pages】:3768-3774
【Authors】: Sheng Li ; Nikos Vlassis ; Jaya Kawale ; Yun Fu
【Abstract】: A widely used method for estimating counterfactuals and causal treatment effects from observational data is nearest-neighbor matching. This typically involves pairing each treated unit with its nearest-in-covariates control unit, and then estimating an average treatment effect from the set of matched pairs. Although straightforward to implement, this estimator is known to suffer from a bias that increases with the dimensionality of the covariate space, which can be undesirable in applications that involve high-dimensional data. To address this problem, we propose a novel estimator that first projects the data to a number of random linear subspaces, and it then estimates the median treatment effect by nearest-neighbor matching in each subspace. We empirically compute the mean square error of the proposed estimator using semi-synthetic data, and we demonstrate the method on real-world digital marketing campaign data. The results show marked improvement over baseline methods.
【Keywords】:
【Paper Link】 【Pages】:3775-3781
【Authors】: Huijie Lin ; Jia Jia ; Liqiang Nie ; Guangyao Shen ; Tat-Seng Chua
【Abstract】: With the rise of social media such as Twitter, people are more willing to convey their stressful life events via these platforms. In a sense, it is feasible to detect stress from social media data for proactive health care. In psychology, stress is composed of stressor and stress level, where stressor further comprises of stressor event and subject. By far, little attention has been paid to estimate exact stressor and stress level from social media data, due to the following challenges: 1) stressor subject identification, 2) stressor event detection, and 3) data collection and representation. To address these problems, we devise a comprehensive scheme to measure a user's stress level from his/her social media data. In particular, we first build a benchmark dataset and extract a rich set of stress-oriented features. We then propose a novel hybrid multi-task model to detect the stressor event and subject, which is capable of modeling the relatedness among stressor events as well as stressor subjects. At last, we lookup an expert-defined stress table with the detected subject and event to estimate the stressor and stress level. Extensive experiments on real-world datasets well verify the effectiveness of our scheme.
【Keywords】:
【Paper Link】 【Pages】:3782-3788
【Authors】: Yang Liu ; Yiling Chen
【Abstract】: In crowdsourcing when there is a lack of verification for contributed answers, output agreement mechanisms are often used to incentivize participants to provide truthful answers when the correct answer is hold by the majority. In this paper, we focus on using output agreement mechanisms to elicit effort, in addition to eliciting truthful answers, from a population of workers. We consider a setting where workers have heterogeneous cost of effort exertion and examine the data requester's problem of deciding the reward level in output agreement for optimal elicitation. In particular, when the requester knows the cost distribution, we derive the optimal reward level for output agreement mechanisms. This is achieved by first characterizing Bayesian Nash equilibria of output agreement mechanisms for a given reward level. When the cost distribution is unknown to the requester, we develop sequential mechanisms that combine learning the cost distribution with incentivizing effort exertion to approximately determine the optimal reward level.
【Keywords】:
【Paper Link】 【Pages】:3789-3796
【Authors】: Yuli Liu ; Yiqun Liu ; Min Zhang ; Shaoping Ma
【Abstract】: A number of existing works have focused on the problem of malicious following activity detection in microblog services. However, most of them make the assumption that the spamming following relationships are either from fraudulent accounts or compromised legitimate users. They therefore developed detection methodologies based on the features derived from this assumption. Recently, a new type of malicious crowdturfing following relationship is provided by the follower market, called voluntary following. Followers who provide voluntary following services (or named volowers) are normal users who are willing to trade their following activities for profit. Since most of their behaviors follow normal patterns, it is difficult for existing methods to detect volowers and their corresponding customers. In this work, we try to solve the voluntary following problem through a newly proposed detection method named DetectVC. This method incorporates both structure information in user following behavior graphs and prior knowledge collected from follower markets. Experimental results on large scale practical microblog data set show that DetectVC is able to detect volowers and their customers simultaneously and it also significantly outperforms existing solutions.
【Keywords】:
【Paper Link】 【Pages】:3797-3803
【Authors】: Chun-Ta Lu ; Sihong Xie ; Weixiang Shao ; Lifang He ; Philip S. Yu
【Abstract】: Nowadays, a large number of new online businesses emerge rapidly. For these emerging businesses, existing recommendation models usually suffer from the data-sparsity. In this paper, we introduce a novel similarity measure, AmpSim (Augmented Meta Path-based Similarity) that takes both the linkage structures and the augmented link attributes into account. By traversing between heterogeneous networks through overlapping entities, AmpSim can easily gather side information from other networks and capture the rich similarity semantics between entities. We further incorporate the similarity information captured by AmpSim in a collective matrix factorization model such that the transferred knowledge can be iteratively propagated across networks to fit the emerging business. Extensive experiments conducted on real-world datasets demonstrate that our method significantly outperforms other state-of-the-art recommendation models in addressing item recommendation for emerging businesses.
【Keywords】:
【Paper Link】 【Pages】:3804-3810
【Authors】: Zhongqi Lu ; Sinno Jialin Pan ; Yong Li ; Jie Jiang ; Qiang Yang
【Abstract】: Accurate user profiling is important for an online recommender system to provide proper personalized recommendations to its users. In many real-world scenarios, the user's interests towards the items may change over time. Therefore, a dynamic and evolutionary user profile is needed. In this work, we come up with a novel evolutionary view of user's profile by proposing a Collaborative Evolution (CE) model, which learns the evolution of user's profiles through the sparse historical data in recommender systems and outputs the prospective user profile of the future. To verify the effectiveness of the proposed model, we conduct experiments on a real-world dataset, which is obtained from the online shopping website of Tencent — www.51buy.com and contains more than 1 million users' shopping records in a time span of more than 180 days. Experimental analyses demonstrate that our proposed CE model can be used to make better future recommendations compared to several state-of-the-art methods.
【Keywords】:
【Paper Link】 【Pages】:3811-3817
【Authors】: Ping Luo ; Ganbin Zhou ; Jiaxi Tang ; Rui Chen ; Zhongjie Yu ; Qing He
【Abstract】: Various hedonic content systems (e.g. mobile apps for video, music, news, jokes, pictures, social networks etc.) increasingly dominate people's daily spare life. This paper studies common regularities of browsing behaviors in these systems, based on a large data set of user logs. We found that despite differences in visit time and user types, the distribution over browsing length for a visit can be described by the inverse Gaussian form with a very high precision. It indicates that the choice threshold model of decision making on continuing browsing or leave does exist. Also, We found that the stimulus intensity, in terms of the amount of recent enjoyed items, affects the probability of continuing browsing in a curve of inverted-U shape. We discuss the possible origin of this curve based on a proposed Award-Aversion Contest model. This hypothesis is supported by the empirical study, which shows that the proposed model can successfully recover the original inverse Gaussian distribution for the browsing length. These browsing regularities can be used to develop better organization of hedonic content, which helps to attract more user dwell time in these systems.
【Keywords】:
【Paper Link】 【Pages】:3818-3824
【Authors】: Jing Ma ; Wei Gao ; Prasenjit Mitra ; Sejeong Kwon ; Bernard J. Jansen ; Kam-Fai Wong ; Meeyoung Cha
【Abstract】: Microblogging platforms are an ideal place for spreading rumors and automatically debunking rumors is a crucial problem. To detect rumors, existing approaches have relied on hand-crafted features for employing machine learning algorithms that require daunting manual effort. Upon facing a dubious claim, people dispute its truthfulness by posting various cues over time, which generates long-distance dependencies of evidence. This paper presents a novel method that learns continuous representations of microblog events for identifying rumors. The proposed model is based on recurrent neural networks (RNN) for learning the hidden representations that capture the variation of contextual information of relevant posts over time. Experimental results on datasets from two real-world microblog platforms demonstrate that (1) the RNN method outperforms state-of-the-art rumor detection models that use hand-crafted features; (2) performance of the RNN-based algorithm is further improved via sophisticated recurrent units and extra hidden layers; (3) RNN-based method detects rumors more quickly and accurately than existing techniques, including the leading online rumor debunking services.
【Keywords】:
【Paper Link】 【Pages】:3825-3831
【Authors】: Shengying Pan ; Kate Larson ; Josh Bradshaw ; Edith Law
【Abstract】: The automation of hiring decisions is a well-studied topic in crowdsourcing. Existing hiring algorithms make a common assumption — that each worker has a level of task competence that is static and does not vary over time. In this work, we explore the question of how to hire workers who can learn over time. Using a medical time series classification task as a case study, we conducted experiments to show that workers' performance does improve with experience and that it is possible to model and predict their learning rate. Furthermore, we propose a dynamic hiring mechanism that accounts for workers' learning potential. Through both simulation and real-world crowdsourcing data, we show that our hiring procedure can lead to high-accuracy outcomes at lower cost compared to other mechanisms.
【Keywords】:
【Paper Link】 【Pages】:3832-3838
【Authors】: Yingwei Pan ; Yehao Li ; Ting Yao ; Tao Mei ; Houqiang Li ; Yong Rui
【Abstract】: Learning video representation is not a trivial task, as video is an information-intensive media where each frame does not exist independently. Locally, a video frame is visually and semantically similar with its adjacent frames. Holistically, a video has its inherent structure — the correlations among video frames. For example, even the frames far from each other may also hold similar semantics. Such context information is therefore important to characterize the intrinsic representation of a video frame. In this paper, we present a novel approach to learn the deep video representation by exploring both local and holistic contexts. Specifically, we propose a triplet sampling mechanism to encode the local temporal relationship of adjacent frames based on their deep representations. In addition, we incorporate the graph structure of the video, as a priori, to holistically preserve the inherent correlations among video frames. Our approach is fully unsupervised and trained in an end-to-end deep convolutional neural network architecture. By extensive experiments, we show that our learned representation can significantly boost several video recognition tasks (retrieval, classification, and highlight detection) over traditional video representations.
【Keywords】:
【Paper Link】 【Pages】:3839-3845
【Authors】: Alexandra Papoutsaki ; Patsorn Sangkloy ; James Laskey ; Nediyana Daskalova ; Jeff Huang ; James Hays
【Abstract】: We introduce WebGazer, an online eye tracker that uses common webcams already present in laptops and mobile devices to infer the eye-gaze locations of web visitors on a page in real time. The eye tracking model self-calibrates by watching web visitors interact with the web page and trains a mapping between features of the eye and positions on the screen. This approach aims to provide a natural experience to everyday users that is not restricted to laboratories and highly controlled user studies. WebGazer has two key components: a pupil detector that can be combined with any eye detection library, and a gaze estimator using regression analysis informed by user interactions. We perform a large remote online study and a small in-person study to evaluate WebGazer. The findings show that WebGazer can learn from user interactions and that its accuracy is sufficient for approximating the user's gaze. As part of this paper, we release the first eye tracking library that can be easily integrated in any website for real-time gaze interactions, usability studies, or web research.
【Keywords】:
【Paper Link】 【Pages】:3846-3853
【Authors】: Yuxin Peng ; Xin Huang ; Jinwei Qi
【Abstract】: Inspired by the progress of deep neural network (DNN) in single-media retrieval, the researchers have applied the DNN to cross-media retrieval. These methods are mainly two-stage learning: the first stage is to generate the separate representation for each media type, and the existing methods only model the intra-media information but ignore the inter-media correlation with the rich complementary context to the intra-media information. The second stage is to get the shared representation by learning the cross-media correlation, and the existing methods learn the shared representation through a shallow network structure, which cannot fully capture the complex cross-media correlation. For addressing the above problems, we propose the cross-media multiple deep network (CMDN) to exploit the complex cross-media correlation by hierarchical learning. In the first stage, CMDN jointly models the intra-media and inter-media information for getting the complementary separate representation of each media type. In the second stage, CMDN hierarchically combines the inter-media and intra-media representations to further learn the rich cross-media correlation by a deeper two-level network strategy, and finally get the shared representation by a stacked network style. Experiment results show that CMDN achieves better performance comparing with several state-of-the-art methods on 3 extensively used cross-media datasets.
【Keywords】:
【Paper Link】 【Pages】:3854-3860
【Authors】: Suvash Sedhain ; Hung Hai Bui ; Jaya Kawale ; Nikos Vlassis ; Branislav Kveton ; Aditya Krishna Menon ; Trung Bui ; Scott Sanner
【Abstract】: Collaborative filtering has emerged as the de facto approach to personalized recommendation problems. However, a scenario that has proven difficult in practice is the one-class collaborative filtering case (OC-CF), where one has examples of items that a user prefers, but no examples of items they do not prefer. In such cases, it is desirable to have recommendation algorithms that are personalized, learning-based, and highly scalable. Existing linear recommenders for OC-CF achieve good performance in benchmarking tasks, but they involve solving a large number of a regression subproblems, limiting their applicability to large-scale problems. We show that it is possible to scale up linear recommenders to big data by learning an OC-CF model in a randomized low-dimensional embedding of the user-item interaction matrix. Our algorithm, Linear-FLow, achieves state-of-the-art performance in a comprehensive set of experiments on standard benchmarks as well as real data.
【Keywords】:
【Paper Link】 【Pages】:3861-3867
【Authors】: Avi Segal ; Ya'akov (Kobi) Gal ; Ece Kamar ; Eric Horvitz ; Alex Bowyer ; Grant Miller
【Abstract】: Volunteer-based crowdsourcing depend critically on maintaining the engagement of participants. We explore a methodology for extending engagement in citizen science by combining machine learning with intervention design. We first present a platform for using real-time predictions about forthcoming disengagement to guide interventions. Then we discuss a set of experiments with delivering different messages to users based on the proximity to the predicted time of disengagement. The messages address motivational factors that were found in prior studies to influence users' engagements. We evaluate this approach on Galaxy Zoo, one of the largest citizen science application on the web, where we traced the behavior and contributions of thousands of users who received intervention messages over a period of a few months. We found sensitivity of the amount of user contributions to both the timing and nature of the message. Specifically, we found that a message emphasizing the helpfulness of individual users significantly increased users' contributions when delivered ac- cording to predicted times of disengagement, but not when delivered at random times. The influence of the message on users' contributions was more pronounced as additional user data was collected and made available to the classifier.
【Keywords】:
【Paper Link】 【Pages】:3868-3874
【Authors】: Chaofeng Sha ; Xiaowei Wu ; Junyu Niu
【Abstract】: The traditional recommendation systems usually aim to improve the recommendation accuracy while overlooking the diversity within the recommended lists. Although some diversification techniques have been designed to recommend top-k items in terms of both relevance and diversity, the coverage of the user's interest is overlooked. In this paper, we propose a general framework to recommend relevant and diverse items which explicitly takes the coverage of user interest into account. Based on the theoretical analysis, we design efficient greedy algorithms to get the near optimal solutions for those NP-hard problems. Experimental results on MovieLens dataset demonstrate that our approach outperforms state-of-the-art techniques in terms of both precision and diversity.
【Keywords】:
【Paper Link】 【Pages】:3875-3881
【Authors】: Yuan Su ; Xi Zhang ; Philip S. Yu ; Wen Hua ; Xiaofang Zhou ; Binxing Fang
【Abstract】: Information diffusion in online social networks has attracted substantial research effort. Although recent models begin to incorporate interactions among contagions, they still don't consider the comprehensive interactions involving users and contagions as a whole. Moreover, the interactions obtained in previous work are modeled as latent factors and thus are difficult to understand and interpret. In this paper, we investigate the contagion adoption behavior by incorporating various types of interactions into a coherent model, and propose a novel interaction-aware diffusion framework called IAD. IAD exploits the social network structures to distinguish user roles, and uses both structures and texts to categorize contagions. Experiments with large-scale Weibo dataset demonstrate that IAD outperforms the state-of-art baselines in terms of F1-score and accuracy, as well as the runtime for learning. In addition, the interactions obtained through learning reveal interesting findings, e.g., food-related contagions have the strongest capability to suppress other contagions' propagation, while advertisement-related contagions have the weakest capability.
【Keywords】:
【Paper Link】 【Pages】:3882-3888
【Authors】: Ryusuke Takahama ; Toshihiro Kamishima ; Hisashi Kashima
【Abstract】: Object ranking is a problem that involves ordering given objects by aggregating pairwise comparison data collected from one or more evaluators; however, the cost of object evaluations is high in some applications. In this paper, we propose an efficient data collection method called progressive comparison, whose objective is to collect many pairwise comparison data while reducing the number of evaluations. We also propose active learning methods to determine which object should be evaluated next in the progressive comparison; we propose two measures of expected model changes, one considering the changes in the evaluation score distributions and the other considering the changes in the winning probabilities. The results of experiments using a synthetic dataset and two real datasets demonstrate that the progressive comparison method achieves high estimation accuracy with a smaller number of evaluations than the standard pairwise comparison method, and that the active learning methods further reduce the number of evaluations as compared with a random sampling method.
【Keywords】:
【Paper Link】 【Pages】:3889-3895
【Authors】: Cunchao Tu ; Weicheng Zhang ; Zhiyuan Liu ; Maosong Sun
【Abstract】: DeepWalk is a typical representation learning method that learns low-dimensional representations for vertices in social networks. Similar to other network representation learning (NRL) models, it encodes the network structure into vertex representations and is learnt in unsupervised form. However, the learnt representations usually lack the ability of discrimination when applied to machine learning tasks, such as vertex classification. In this paper, we overcome this challenge by proposing a novel semi-supervised model, max-margin DeepWalk (MMDW). MMDW is a unified NRL framework that jointly optimizes the max-margin classifier and the aimed social representation learning model. Influenced by the max-margin classifier, the learnt representations not only contain the network structure, but also have the characteristic of discrimination. The visualizations of learnt representations indicate that our model is more discriminative than unsupervised ones, and the experimental results on vertex classification demonstrate that our method achieves a significant improvement than other state-of-the-art methods.
【Keywords】:
【Paper Link】 【Pages】:3896-3902
【Authors】: Jacopo Urbani ; Sourav Dutta ; Sairam Gurajada ; Gerhard Weikum
【Abstract】: Many Web applications require efficient querying of large Knowledge Graphs (KGs). We propose KOGNAC, a dictionary-encoding algorithm designed to improve SPARQL querying with a judicious combination of statistical and semantic techniques. In KOGNAC, frequent terms are detected with a frequency approximation algorithm and encoded to maximise compression. Infrequent terms are semantically grouped into ontological classes and encoded to increase data locality. We evaluated KOGNAC in combination with state-of-the-art RDF engines, and observed that it significantly improves SPARQL querying on KGs with up to 1B edges.
【Keywords】:
【Paper Link】 【Pages】:3903-3909
【Authors】: Jingjing Wang ; Changsung Kang ; Yi Chang ; Jiawei Han
【Abstract】: Hostnames such as en.wikipedia.org and www.amazon.com are strong indicators of the content they host. The relevant hostnames for a query can be a signature that captures the query intent. In this study, we learn the hostname preference of queries, which are further utilized to enhance search relevance. Implicit and explicit query intent are modeled simultaneously by a feature aware matrix completion framework. A block-wise parallel algorithm was developed on top of the Spark MLlib for fast optimization of feature aware matrix completion. The optimization completes within minutes at the scale of a million x million matrix, which enables efficient experimental studies at the web scale. Evaluation of the learned hostname preference is performed both intrinsically on test errors, and extrinsically on the impact on search ranking relevance. Experimental results demonstrate that capturing hostname preference can significantly boost the retrieval performance.
【Keywords】:
【Paper Link】 【Pages】:3910-3916
【Authors】: Keqiang Wang ; Wayne Xin Zhao ; Hongwei Peng ; Xiaoling Wang
【Abstract】: Recently, Local Matrix Factorization (LMF) has been shown to be more effective than traditional matrix factorization for rating prediction. The core idea for LMF is to first partition the original matrix into several smaller submatrices, further exploit local structures of submatrices for better low-rank approximation. Various clustering-based methods with heuristic extensions have been proposed for LMF in the literature. To develop a more principled solution for LMF, this paper presents a Bayesian Probabilistic Multi-Topic Matrix Factorization model. We treat the set of the rated items by a useras a document, and employ latent topic models to cluster items as topics. Subsequently, a user has a distribution over the set of topics. We further set topic-specific latent vectors for both users and items. The final prediction is obtained by an ensemble of the results from the corresponding topic-specific latent vectorsin each topic. Using a multi-topic latent representation, our model is more powerful to reflect the complex characteristics for users and items in rating prediction, and enhance the model interpretability. Extensive experiments on large real-world datasets demonstrate the effectiveness of the proposed model.
【Keywords】:
【Paper Link】 【Pages】:3917-3923
【Authors】: Zheng Wang ; Chaokun Wang ; Jisheng Pei ; Xiaojun Ye ; Philip S. Yu
【Abstract】: In social network sites (SNS), propagation histories which record the information diffusion process can be used to explain to users what happened in their networks. However, these histories easily grow in size and complexity, limiting their intuitive understanding by users. To reduce this information overload, in this paper, we present the problem of propagation history ranking. The goal is to rank participant edges/nodes by their contribution to the diffusion. Firstly, we discuss and adapt Difference of Causal Effects (DCE) as the ranking criterion. Then, to avoid the complex calculation of DCE, we propose a resp-cap ranking strategy by adopting two indicators. The first is responsibility which captures the necessary face of causal effects. We further give an approximate algorithm for this indicator. The second is capability which is defined to capture the sufficient face of causal effects. Finally, promising experimental results are presented to verify the feasibility of our method.
【Keywords】:
【Paper Link】 【Pages】:3924-3930
【Authors】: Bo Xu ; Chenhao Xie ; Yi Zhang ; Yanghua Xiao ; Haixun Wang ; Wei Wang
【Abstract】: Categories play a fundamental role in human cognition. Defining features (short for DFs) are the key elements to define a category, which enables machines to categorize objects. Categories enriched with their DFs significantly improve the machine's ability of categorization and benefit many applications built upon categorization. However, defining features can rarely be found for categories in current knowledge bases. Traditional efforts such as manual construction by domain experts are not practical to find defining features for millions of categories. In this paper, we make the first attempt to automatically find defining features for millions of categories in the real world. We formalize the defining feature learning problem and propose a bootstrapping solution to learn defining features from the features of entities belonging to a category. Experimental results show the effectiveness and efficiency of our method. Finally, we find defining features for overall 60,247 categories with acceptable accuracy.
【Keywords】:
【Paper Link】 【Pages】:3931-3937
【Authors】: Ting Yao ; Fuchen Long ; Tao Mei ; Yong Rui
【Abstract】: Hashing techniques have been intensively investigated for large scale vision applications. Recent research has shown that leveraging supervised information can lead to high quality hashing. However, most existing supervised hashing methods only construct similarity-preserving hash codes. Observing that semantic structures carry complementary information, we propose the idea of co-training for hashing, by jointly learning projections from image representations to hash codes and classification. Specifically, a novel deep semantic-preserving and ranking-based hashing (DSRH) architecture is presented, which consists of three components: a deep CNN for learning image representations, a hash stream of a binary mapping layer by evenly dividing the learnt representations into multiple bags and encoding each bag into one hash bit, and a classification stream. Meanwhile, our model is learnt under two constraints at the top loss layer of hash stream: a triplet ranking loss and orthogonality constraint. The former aims to preserve the relative similarity ordering in the triplets, while the latter makes different hash bit as independent as possible. We have conducted experiments on CIFAR-10 and NUS-WIDE image benchmarks, demonstrating that our approach can provide superior image search accuracy than other state-of-the-art hashing techniques.
【Keywords】:
【Paper Link】 【Pages】:3938-3944
【Authors】: Hongyi Zhang ; Tong Zhao ; Irwin King ; Michael R. Lyu
【Abstract】: Overlapping community detection has drawn much attention recently since it allows nodes in a network to have multiple community memberships. A standard framework to deal with overlapping community detection is Matrix Factorization (MF). Although all existing MF-based approaches use links as input to identify communities, the relationship between links and communities is still under-investigated. Most of the approaches only view links as consequences of communities (community-to-link) but fail to explore how nodes' community memberships can be represented by their linked neighbors (link-to-community). In this paper, we propose a Homophily-based Nonnegative Matrix Factorization (HNMF) to model both-sided relationships between links and communities. From the community-to-link perspective, we apply a preference-based pairwise function by assuming that nodes with common communities have a higher probability to build links than those without common communities. From the link-to-community perspective, we propose a new community representation learning with network embedding by assuming that linked nodes have similar community representations. We conduct experiments on several real-world networks and the results show that our HNMF model is able to find communities with better quality compared with state-of-the-art baselines.
【Keywords】:
【Paper Link】 【Pages】:3945-3951
【Authors】: Lili Zhao ; Zhongqi Lu ; Sinno Jialin Pan ; Qiang Yang
【Abstract】: We present a novel model for movie recommendations using additional visual features extracted from pictorial data like posters and still frames, to better understand movies. In particular, several context-based methods for recommendation are shown to be special cases of our proposed framework. Unlike existing context-based approaches, our method can be used to incorporate visual features — features that are lacking in existing context-based approaches for movie recommendations. In reality, movie posters and still frames provide us with rich knowledge for understanding movies, users' preferences as well. For instance, user may want to watch a movie at the minute when she/he finds some released posters or still frames attractive. Unfortunately, such unique features cannot be revealed from rating data or other form of context that being used in most of existing methods. In this paper, we take a step in this direction and investigate both low-level and high-level visual features from the movie posters and still frames for further improvement of recommendation methods. A comprehensive set of experiments on real world datasets shows that our approach leads to significant improvement over the state-of-the-art methods.
【Keywords】:
【Paper Link】 【Pages】:3952-3958
【Authors】: Haoti Zhong ; Hao Li ; Anna Cinzia Squicciarini ; Sarah Michele Rajtmajer ; Christopher Griffin ; David J. Miller ; Cornelia Caragea
【Abstract】: We study detection of cyberbullying in photo-sharing networks, with an eye on developing early warning mechanisms for the prediction of posted images vulnerable to attacks. Given the overwhelming increase in media accompanying text in online social networks, we investigate use of posted images and captions for improved detection of bullying in response to shared content. We validate our approaches on a dataset of over 3000 images along with peer-generated comments posted on the Instagram photo-sharing network, running comprehensive experiments using a variety of classifiers and feature sets. In addition to standard image and text features, we leverage several novel features including topics determined from image captions and a pretrained convolutional neural network on image pixels. We identify the importance of these advanced features in assisting detection of cyberbullying in posted comments. We also provide results on classification of images and captions themselves as potential targets for cyberbullies.
【Keywords】:
【Paper Link】 【Pages】:3959-3967
【Authors】: Lei Zhu ; Jialie Shen ; Xiaobai Liu ; Liang Xie ; Liqiang Nie
【Abstract】: Mobile Landmark Search (MLS) recently receives increasing attention. However, it still remains unsolved due to two important issues. One is high bandwidth consumption of query transmission, and the other is the huge visual variations of query images. This paper proposes a Canonical View based Compact Visual Representation (2CVR) to handle these problems via novel three stage learning. First, a submodular function is designed to measure visual representativeness and redundancy of a view set. With it, canonical views, which capture key visual appearances of landmark with limited redundancy, are efficiently discovered with an iterative mining strategy. Second, multimodal sparse coding is applied to transform multiple visual features into an intermediate representation which can robustly characterize visual contents of varied landmark images with only fixed canonical views. Finally, compact binary codes are learned on intermediate representation within a tailored binary embedding model which preserves visual relations of images measured with canonical views and removes noises. With 2CVR, robust visual query processing, low cost of query transmission, and fast search process are simultaneously supported. Experiments demonstrate the superior performance of 2CVR over several state-of-the-art methods.
【Keywords】:
【Paper Link】 【Pages】:3968-3969
【Authors】: Zohreh Alavi
【Abstract】: The ability to act and respond to exogenous events in dynamic environments is crucial for robust autonomy. In dynamic environments, external changes may occur that prevent an agent from reaching its goal(s). I am interested in the design of reasoning and planning components operating in environments that undergo changes in real time. My goal is to develop a framework for fully integrated planning, execution and vision in dynamic environments.
【Keywords】:
【Paper Link】 【Pages】:3970-3971
【Authors】: Martin Aleksandrov
【Abstract】: Hunger is a major problem worldwide. Food banks around the globe combine forces with various welfare agencies towards alleviating the hunger by assisting people in need. For example, Foodbank Australia cooperates with local charities in order to effectively allocate food as it is donated. In 2014, nearly 10% of these relief organizations could not meet the demand and thus left around 24,000 children with no breakfast in their schools. Can we improve the food allocation? Further, the Foodbanking network in Canada has a long-standing tradition in handling customer demands, but in the last year 60% of their sponsorship covered the delivery of the food. Can we reduce the transportation costs implied by the food allocation? Finally, the Meal Gap in New York reached 250 millions in 2014. How do we allocate food in cities that "never sleep" and in which there are high time and spatial dynamics? Evidently, a food bank needs an allocation mechanism that takes all these features into account. Such a mechanism should be able to (1) allocate resources online, (2) be robust to stochastic changes in the allocation preferences and (3) inform dispatching solutions. I address exactly such complex real-world features in here.
【Keywords】:
【Paper Link】 【Pages】:3972-3973
【Authors】: Ankit Anand
【Abstract】: Many traditional AI algorithms fail to scale as the size of state space increases exponentially with the number of features. One way to reduce computation in such scenarios is to reduce the problem size by grouping symmetric states together and then running the algorithm on the reduced problem. The focus of this work is to exploit symmetry in problems of sequential decision making and probabilistic inference. Our recent work- ASAP-UCT defines new State-Action Pair (SAP) symmetries in Markov Decision Processes. We also apply these SAP symmetries in Monte Carlo Tree Search (MCTS) framework. In probabilistic inference, we expand the notion of unconditional symmetries to contextual symmetries and apply them in Markov Chain Monte Carlo (MCMC) methods. In future, we plan to explore interesting links in symmetry exploitation in different problems and aim to develop a generic symmetry based framework.
【Keywords】:
【Paper Link】 【Pages】:3974-3975
【Authors】: Evgenii Balai
【Abstract】: My research focuses on investigation and improvement of knowledge representation (KR) language P-log which was designed to reason about both logical and probabilistic knowledge. In particular, I aim to extend P-log with new constructs, clarify its semantics, develop a new efficient inference engine for it and establish its relationship with other related formalisms. Successful completion of this work will greatly increase the scope of practical applications of the language.
【Keywords】:
【Paper Link】 【Pages】:3976-3977
【Authors】: Alberto Camacho
【Abstract】: In the last decade, we have seen an exponential increase in the number of devices connected to the Internet, with a commensurate explosion in the availability of data. New applications such as those related to smart cities exemplify the need for principled techniques for automated intelligent decision making based on available data. Many decision-making problems require reasoning in large and complex state spaces, sometimes under stringent time constraints. The nature of these problems suggests that planning approaches could be used to find solutions efficiently. Automated planning is the basis for addressing a diversity of problems beyond classical planning such as automated diagnosis, controller synthesis, and story understanding. Nevertheless, many planning paradigms make assumptions that do not hold in real-world settings. Our work focuses on exploring planning paradigms that capture properties of real-world decision-making applications. These properties include the ability to model nondeterminism in the outcome of actions, the ability to deal with complex objectives that are temporally extended (in contrast to final-state goals) some of which may be necessary and other simply desirable to optimize for. Finally, we are interested in dealing with incomplete information. Addressing this class of problems presents challenges related to problem specification, modeling, and computationally efficient techniques for generating solutions.
【Keywords】:
【Paper Link】 【Pages】:3978-3979
【Authors】: Liron Cohen ; Sven Koenig
【Abstract】: The multi-agent path finding (MAPF) problem is defined as follows: Given a graph and a set of agents with unique start and goal vertices, find collision-free paths for all agents from their respective start vertices to their respective goal vertices. Our objective is to minimize the the total arrival time. MAPF has many applications such as video games, traffic control and robotics.
【Keywords】:
【Paper Link】 【Pages】:3980-3981
【Authors】: Andrew Cropper
【Abstract】: Inductive programming approaches typically rely on an Occamist bias to select hypotheses with minimal textual complexity. This approach, however, fails to distinguish between the efficiencies of hypothesised programs, such as merge sort (O(n log n)) and bubble sort (O(n2)). We address this issue by introducing techniques to learn logic programs with minimal resource complexity. We describe an algorithm proven to learn minimal resource complexity robot strategies, and we propose future work to generalise the approach to a broader class of programs.
【Keywords】:
【Paper Link】 【Pages】:3982-3983
【Authors】: Felipe Leno da Silva ; Anna Helena Reali Costa
【Abstract】: Reinforcement learning methods have successfully been applied to build autonomous agents that solve many sequential decision making problems. However, agents need a long time to learn a suitable policy, specially when multiple autonomous agents are in the environment. This research aims to propose a Transfer Learning (TL) framework to accelerate learning by exploiting two knowledge sources: (i) previously learned tasks; and (ii) advising from a more experienced agent. The definition of such framework requires answering several challenging research questions, including: How to abstract and represent knowledge, in order to allow generalization and posterior reuse?, How and when to transfer and receive knowledge in an efficient manner?, and How to evaluate the transfer quality in a Multiagent scenario?.
【Keywords】:
【Paper Link】 【Pages】:3984-3985
【Authors】: Steven Damer
【Abstract】: My primary research interest is social behavior for software agents to achieve cooperation in general-sum normal form games. An agent can easily be programmed to constantly cooperate in a normal form game, but such an agent is not suitable for environments with potentially hostile opponents. For a rational agent, the main reason to cooperate is to induce reciprocation; to reciprocate it is necessary to determine which moves are cooperative. In constant-sum games cooperation is impossible because any gain by one agent isa loss by the other agent. In other games (such as Prisoner's Dilemma) it is easy to identify cooperative moves because the opponent's payoffs for that move strictly dominate the other moves of the agent. In general it is not easy to identify cooperative strategies in arbitrary general-sum games.
【Keywords】:
【Paper Link】 【Pages】:3986-3987
【Authors】: Dustin Dannenhauer
【Abstract】: As intelligent systems become a regular part of our everyday lives, robust and safe operation is ever more important. My research focus is to endow agents with the ability to monitor themselves in order to detect when their behavior has exceeded their boundaries. Previously, we have explored different forms of expectations for anomaly detection in agents operating in Real-Time Strategy (RTS) games, as well as dynamic domains involving planning and execution. My current work aims to achieve agents that can reason about and use expectations in both dynamic and partially observable domains, as well as investigating meta-cognitive expectations for detecting anomalies in the agent's own cognitive processes (reasoning, planning, etc) instead of anomalies in the world.
【Keywords】:
【Paper Link】 【Pages】:3988-3989
【Authors】: Claudia Pérez-D'Arpino ; Julie A. Shah
【Abstract】: The efficient and safe performance of collaborative robots requires advancements in perception, control, design and algorithms, among other factors. With regard to algorithms, representing the structure of collaborative tasks and reasoning about progress toward task completion in an on-line fashion enables a robot to be a fluent and safe collaborator based on its ability to predict the next actions of a human agent. With this goal in mind, we focus on real-time target prediction of human reaching motion and present an algorithm based on time series classification. Results from on-line testing involving a tabletop task with a PR2 robot yielded 70% prediction accuracy with 400 msec of observed trajectory.
【Keywords】:
【Paper Link】 【Pages】:3990-3991
【Authors】: Negar Ghourchian
【Abstract】: We consider the problem of analyzing people's mobility and movement patterns from their location history, gathered by mobile devices. Human mobility traces can be extremely complex and unpredictable, by nature, which makes it hard to construct accurate models of mobility behavior. In this work, we present a novel high-level strategy for mobility data analysis based on Hierarchical Dirichlet process, which is a powerful probabilistic model for clustering grouped data. We evaluate our unsupervised approach on two real-world datasets.
【Keywords】:
【Paper Link】 【Pages】:3992-3993
【Authors】: Julio Godoy
【Abstract】: In real-time multi-agent navigation, agents need to move towards their goal positions while adapting their paths to avoid potential collisions with other agents and static obstacles. Existing methods compute motions that are optimal locally but do not account for the motions of the other agents, producing inefficient global motions especially when many agents move in a crowded space. In my thesis work, each agent has only a limited sensing range and uses online action selection techniques to dynamically adapt its motion to the local conditions. Experimental results obtained in simulation under different conditions show that the agents reach their destinations faster and use motions that minimize their overall energy consumption.
【Keywords】:
【Paper Link】 【Pages】:3994-3995
【Authors】: Jasmin Grosinger
【Abstract】: In order to be proactive, robots should be capable of generating and selecting their own goals, and pursuing activities towards their achievement. Goal reasoning has focused on the former set of cognitive abilities, and automated planning on the latter. Despite the existence of robots that possess both capabilities, we lack a general understanding of how to combine goal generation and goal achievement. In my work I introduce the notion of equilibrium maintenance as a contribution to this understanding. Formal evidence is provided that equilibrium maintenance is conducive to proactive robots, and I demonstrate my approach in a closed loop with a real robot in a smart home.
【Keywords】:
【Paper Link】 【Pages】:3996-3997
【Authors】: Ping Hou
【Abstract】: While probabilistic planning have been extensively studied by artificial intelligence communities for planning under uncertainty, the objective to minimize the expected cumulative cost is inappropriate for high-stake planning problems. With this motivation in mind, we revisit the Risk-Sensitive criterion (RS-criterion), where the objective is to find a policy that maximizes the probability that the cumulative cost is within some user-defined cost threshold. By combining goal-directed MDPs and POMDPs with the RS-criterion, the corresponding risk-sensitive probabilistic planning models —Risk-Sensitive MDPs (RS-MDPs) and Risk-Sensitive POMDPs (RS-POMDPs) — can be formalized. The overall scope of this research is to develop efficient and scalable RS-MDP and RS-POMDP algorithms.
【Keywords】:
【Paper Link】 【Pages】:3998-3999
【Authors】: Aaron Isaksen
【Abstract】: Artificial intelligence can model and predict how humans will react to game systems. Using models of human-like imperfect play, one can estimate how quantitative changes to a game will impact a player's qualitative experience. I discuss my research to use artificial intelligence to automatically play, analyze, and design certain classes of games. Not only does it generate new content quickly and effectively, this research also provides some insights into why humans find some games more difficult than others, and how player skill measurably improves over time.
【Keywords】:
【Paper Link】 【Pages】:4000-4001
【Authors】: Bilal Kartal
【Abstract】: Multi-agent planning approaches are employed for many problems including task allocation, surveillance and video games. In the first part of my thesis, we study two multi-robot planning problems, i.e. patrolling and task allocation. For the patrolling problem, we present a novel stochastic search technique, Monte Carlo Tree Search with Useful Cycles, that can generate optimal cyclic patrol policies with theoretical convergence guarantees. For the multi-robot task allocation problem, we propose an Monte Carlo Tree Search based satisficing method using branch and bound paradigm along with a novel search parallelization technique. In the second part of my thesis, we develop a stochastic multi-agent narrative planner employing MCTS along with new heuristic and pruning methods applicable for other planning domains as well.
【Keywords】:
【Paper Link】 【Pages】:4002-4003
【Authors】: Chris Kedzie ; Kathleen McKeown
【Abstract】: During crises, information is critical for responders and victims. When the event is significant, as in the case of hurricane Sandy, the amount of content produced by traditional news outlets, relief organizations, and social media vastly overwhelms those trying to monitor the situation. An emerging task in this space is to monitor an event as it unfolds over time by processing an associated stream of documents to produce a rolling update summary containing the most salient information with respect to the event. In this thesis, we develop two extractive summarization systems for streaming text data. Both systems explicitly predict the salience of input stream text to create a rolling summary. Finally, we discuss our proposed work for combining these systems with an abstractive text generation model.
【Keywords】:
【Paper Link】 【Pages】:4004-4005
【Authors】: Elias B. Khalil
【Abstract】: Mixed Integer Programs (MIP) are solved exactly by tree-based branch-and-bound search. However, various components of the algorithm involve making decisions that are currently addressed heuristically. Instead, I propose to use machine learning (ML) approaches such as supervised ranking and multi-armed bandits to make better-informed, input-specific decisions during MIP branch-and-bound. My thesis aims at improving the overall performance of MIP solvers. To illustrate the potential for ML in MIP, I have so far tackled branching variable selection, a crucial component of the search procedure, showing that ML approaches for variable selection can outperform traditional heuristics.
【Keywords】:
【Paper Link】 【Pages】:4006-4007
【Authors】: Faiza Khan Khattak
【Abstract】: One of the main challenges in crowd-labeling is to control for or determine in advance the proportion of low-quality/malicious labelers. We propose methods that estimate the labeler and data instance related parameters using frequentist and Bayesian approaches. All these approaches are based on expert-labeled instance (ground truth) for a small percentage of data to learn the parameters. We also derive a lower bound on the number of expert-labeled instances needed to get better quality labels.
【Keywords】:
【Paper Link】 【Pages】:4008-4009
【Authors】: Caner Komurlu
【Abstract】: In supervised learning, many techniques focus on optimizing training phase to increase prediction performance. Active inference, a relatively novel paradigm, aims to decrease overall prediction error via selective collection of some labels based on relations among instances. In this research, we use dynamic Bayesian networks to model temporal systems and we apply active inference to dynamically choose variables for observation so as to improve prediction on unobserved variables
【Keywords】:
【Paper Link】 【Pages】:4010-4011
【Authors】: Sheng Li
【Abstract】: Learning compact representations from high-dimensional and large-scale data plays an essential role in many real-world applications. However, many existing methods show limited performance when data are contaminated with severe noise. To address this challenge, we have proposed several effective methods to extract robust data representations, such as balanced graphs, discriminative subspaces, and robust dictionaries. In addition, several topics are provided as future work.
【Keywords】:
【Paper Link】 【Pages】:4012-4013
【Authors】: Valentin Mayer-Eichberger
【Abstract】: Boolean Satisfiability (SAT) solvers are a mature technology to solve hard combinatorial problems. The input to a SAT solver is the problem translated to propositional logic in conjunctive normal form (CNF). This thesis studies such translations and aims to make SAT solvers more accessible to non-encoding experts.
【Keywords】:
【Paper Link】 【Pages】:4014-4015
【Authors】: Ciaran McCreesh
【Abstract】: We look at problems involving finding subgraphs in larger graphs, such as the maximum clique problem, the subgraph isomorphism problem, and the maximum common subgraph problem. We investigate variable and value ordering heuristics, different inference strategies, intelligent backtracking search (backjumping), and bit- and thread-parallelism to exploit modern hardware.
【Keywords】:
【Paper Link】 【Pages】:4016-4017
【Authors】: Gulnar Mehdi ; Sebastian Brandt ; Mikhail Roshchin ; Thomas A. Runkler
【Abstract】: Massive data streams from sensors and devices are prominent form of industrial data generated during condition-monitoring and diagnosis of complex systems. Data analytics and reasoning has emerged as a vital tool to harness massive data sets, providing insights into historical and real-time system conditions; enhanced decision support, reliability and cost reduction. However, application of data analytics is mainly challenged by the complexity of data-access, integration, domain-specific query support and contextual reasoning capabilities. The current state-of-the-art only uses dedicated scenarios and sensors, but this limits reuse, scalability and are not sufficient for an integrated solution. Our thesis investigates if semantic technology can be a potential solution to interact and leverage data analytics for operational use. First, we have studied related work and utilized ontology-based data access (OBDA) techniques for semantic interpretation of diagnosis for Siemens Turbine use-case. Secondly, we have extended our solution to support any analytical workflow by means of an ontology.
【Keywords】:
【Paper Link】 【Pages】:4018-4019
【Authors】: Fei Mi ; Boi Faltings
【Abstract】: Machine learning is often used to acquire knowledge in domains that undergo frequent changes, such as networks, social media, or markets. These frequent changes poses a challenge to most machine learning methods as they have difficulty adapting. So my thesis topic focus on adaptive machine learning models. At the first step, we consider a forum content recommender system for massive open online courses (MOOCs) as an example of an application where recommendations have to adapt to new items and evolving user preferences. We formalize the recommendation problem as a sequence prediction problem and compare different recommendation methods, including a new method called context tree (CT). The results show that the CT recommender performs much better than other methods. We analyze the reasons for this and demonstrate that it is because of better adaptation to changes in the domain.
【Keywords】:
【Paper Link】 【Pages】:4020-4021
【Authors】: Decebal Constantin Mocanu
【Abstract】: Traditionally science is done using the reductionism paradigm. Artificial intelligence does not make an exception and it follows the same strategy. At the same time, network science tries to study complex systems as a whole. This Ph.D. research takes an alternative approach to the reductionism strategy, and tries to advance both fields, i.e. artificial intelligence and network science, by searching for the synergy between them, while not ignoring any other source of inspiration, e.g. neuroscience.
【Keywords】:
【Paper Link】 【Pages】:4022-4023
【Authors】: Banafsheh Rekabdar
【Abstract】: Learning and recognizing spatio-temporal patterns is an important problem for all biological systems. Gestures, movements and activities, all encompass both spatial and temporal information that is critical for implicit communication and learning. This paper presents a novel, unsupervised approach for learning, recognizing and early classifying spatio-temporal patterns using spiking neural networks for human robotic domains. The proposed spiking approach has four variations which have been validated on images of handwritten digits and human hand gestures and motions. The main contributions of this work are as follows: i) it requires a very small number of training examples, ii) it enables early recognition from only partial information of the pattern, iii) it learns patterns in an unsupervised manner, iv) it accepts variable sized input patterns, v) it is invariant to scale and translation, vi) it can recognize patterns in real-time and, vii) it is suitable for human-robot interaction applications and has been successfully tested on a PR2 robot. We also compared all variations of this approach with well-known supervised machine learning methods including support vector machines (SVM), regularized logistic regression (LR) and ensemble neural networks (ENN). Although our approach is unsupervised, it outperforms others and in some cases, provides comparable results with other methods.
【Keywords】:
【Paper Link】 【Pages】:4024-4025
【Authors】: Zeynep Gözen Saribatur
【Abstract】: As autonomous systems become more common in our lives, the issue of verifying that they behave as intended and that their design policies are correct becomes more important. This thesis aims to build foundations for such a verification capability for policies with a reactive behavior, with a focus on combining the representation power of action languages with model checking techniques.
【Keywords】:
【Paper Link】 【Pages】:4026-4027
【Authors】: Arpit Sharma
【Abstract】: There are various aspects of making computers understand natural language. Semantic parsing and reasoning on commonsense knowledge are the two important ones. Many NLU tasks such as question answering and co-reference resolution require semantic parsing of text and reasoning with different kinds of commonsense knowledge. In this work we present our progress towards these milestones of NLU. We demonstrate the steps we took towards the goal and the tools/techniques we developed, such as a semantic parser and a novel algorithm to automatically acquire commonsense knowledge from text. We also show the usefulness of the developed tools by applying them to solve tasks such as hard co-reference resolution. This is an ongoing research and in this paper we present our current progress and the future plans to reach the goal of developing a fully autonomous natural language understanding framework.
【Keywords】:
【Paper Link】 【Pages】:4028-4029
【Authors】: Sam Snodgrass
【Abstract】: Procedural content generation (PCG) studies the algorithmic creation of content (e.g., textures, maps), typically for games. PCG has become a popular research topic in recent years, but little has been done in terms of generalized content generators: approaches that can generate content for a variety of games without hand-tuning. We are interested in exploring statistical algorithms that could lead to generalized procedural map generators.
【Keywords】:
【Paper Link】 【Pages】:4030-4031
【Authors】: Xingyu Su
【Abstract】: This extended abstract outlines my PhD research, which addresses the problem of on-line model-based diagnosis of Discrete Event Systems (DES). A DES model represents state dynamics in a discrete manner. Given a flow of observable events generated by a DES model, diagnosis aims at deciding whether a system is running normally or is experiencing faulty behaviors. The main challenge is to deal with the complexity of a diagnosis problem, which has to monitor an observation flow on the fly and generate a succession of the states that the system is possibly in, called belief state. Previous work has proposed exact diagnosis, which attempts to compute a belief state at any time consistent with the observation flow from the time when the system starts operating to the current time. The main drawback is the inability to follow the observation flow for a large system because the size of each belief state has been proved to be exponential in the number of system states. Furthermore, the temporal complexity to handle the exact belief states remains a problem. Because diagnosis of DES is a hard problem, the use of faster diagnostic algorithms that do not perform an exact diagnosis is often inevitable. However, those algorithms may not be as precise as an exact model-based diagnostic algorithm to diagnose a diagnosable system. This work has four contributions. First, it proposes to verify the precision of an imprecise diagnostic algorithm w.r.t. a diagnosable DES model by constructing a simulation, which is a finite state machine that represents how a diagnostic algorithm works for a DES model. Second, this work proposes window-based diagnostic algorithms, called Independent-Window Algorithms (IWAs). IWAs only diagnose on the very last events of the observation flow and forget about the past. Third, this work proposes a compromise between the two extreme strategies of exact diagnosis and IWAs by looking for the minimum piece of information to remember from the past so that a window-based algorithm ensures the same precision as using the exact diagnosis. This work proposes Time-Window Algorithms (TWAs), which are extensions to IWAs. TWAs carry over some information about the current system state from one time window to the next. Fourth, this work evaluates IWAs and TWAs through experiments and compares their performance with the exact diagnosis encoded by Binary Decision Diagrams. This work also examines the impact of the time window selections on the performance of IWAs and TWAs.
【Keywords】:
【Paper Link】 【Pages】:4032-4033
【Authors】: Antonela Tommasel
【Abstract】: Short-texts accentuate the challenges posed by the high feature space dimensionality of text learning tasks. The linked nature of social data causes new dimensions to be added to the feature space, which, also becomes sparser. Thus, efficient and scalable online feature selection becomes a crucial requirement of numerous large-scale social applications. This thesis proposes an online feature selection technique for high-dimensional data based on both social and content-based information for the real-time classification of short-text streams coming from social media. The main objective of this thesis is to define and evaluate a new intelligent text mining technique for enhancing the process of knowledge discovery in social-media. This technique would help in the development of new and more effective models for personalisation and recommendation of content in social environments.
【Keywords】:
【Paper Link】 【Pages】:4034-4035
【Authors】: Diana Troanca
【Abstract】: Conceptual knowledge is closely related to a deeper understanding of existing facts and relationships, but also to the argumentation and communication of why something happens in a particular way. Formal Concept Analysis (FCA) is the core of Conceptual Knowledge Processing. It emerged from applied mathematics and quickly developed into a powerful framework for knowledge representation. It is based on a set theoretical semantics and provides a rich amount of mathematical instruments for representation, acquiring, retrieval, discovery and further processing of knowledge. FCA was introduced in the dyadic setting and extended to a triadic and eventually n-adic setting. Intuitively, dyadic datasets can be understood as objects related to attributes and, in addition, to conditions for the triadic case. FCA defines concepts as maximal clusters of data in which all elements are mutually interrelated. A common problem for n-adic FCA is concept visualization and navigation. The goal of my thesis is to find visualization and navigation paradigms that can be applied to higher-dimensional datasets. Therefore, we study the triadic case and propose several visualization and navigational approaches. Furthermore, we evaluate these approaches, study their generalizations and extend them, where possible, to n-ary formal contexts.
【Keywords】:
【Paper Link】 【Pages】:4036-4037
【Authors】: Josep Valls-Vargas
【Abstract】: Our research focuses on the problem of automatically acquiring structured narrative information from natural language. We have focused on character extraction and narrative role identification from a corpus of Slavic folktales. To address natural language processing (NLP) issues in this particular domain we have explored alternatives to linear pipelined architectures for information extraction, specifically the idea of feedback loops that allow feeding information produced by later modules of the pipeline back to earlier modules. We propose the use of domain knowledge to improve core NLP tasks and the overall performance of our system.
【Keywords】:
【Paper Link】 【Pages】:4038-4039
【Authors】: Przemyslaw Andrzej Walega
【Abstract】: The aim of my work is to establish a computational framework for commonsense spatial reasoning about dynamic domains. The work accomplished so far consists of theoretical investigation of a framework based on a paradigm of Answer Set Programming Modulo Theories and its implementation. The developed system enables to integrate geometrical and qualitative spatial information, reason about indirect spatial effects and perform non-monotonic reasoning in a context of spatio-temporal contexts. In future it might be applied to a wide range of dynamic domains such as cognitive robotics, computer-aided architecture design, geographic information systems, etc.
【Keywords】:
【Paper Link】 【Pages】:4040-4041
【Authors】: John Winder
【Abstract】: I am developing a framework for anomaly reasoning for agents that plan and learn in complex, sometimes unfamiliar domains. Anomaly reasoning encompasses recognizing, interpreting, and reacting to unfamiliar objects or familiar objects appearing in unexpected contexts. As a first approach, I propose an interpretation method in which agents form concepts from perceptions to create new representations for use in planning and decision making. An anomaly reasoning framework will make agents more versatile, facilitate learning transfer by pruning irrelevant features, relate new to known phenomena with appropriate similarity metrics, and guide an agent to aspects of the environment most significant to its goals.
【Keywords】:
【Paper Link】 【Pages】:4042-4043
【Authors】: David R. Winer
【Abstract】: Typical narrative generation systems adopt a story-then-discourse pipeline approach in which fabula (i.e. story) is procedurally generated and provided as input for generating discourse and narration (e.g. text or animation). However, stories produced independently from a communicative plan are not guaranteed to have properties which are readily tellable or worth telling. This extended abstract describes an approach to narrative planning in which constraints for a story are discovered as part of the search for compatible story and discourse solutions.
【Keywords】:
【Paper Link】 【Pages】:4044-4045
【Authors】: Daqing Yi
【Abstract】: Understanding qualitative instructions may enable robots to work with untrained human users. We propose a framework supporting a robot that plans a path using information specified by a human using natural language; the path may contain multiple criteria and topological requirements. The framework supports the conversion from a qualitative sentence to a quantitative path for robotic navigation.
【Keywords】:
【Paper Link】 【Pages】:4046-4049
【Authors】: Dingwen Zhang
【Abstract】: Because of playing one of the most important roles in the artificial intelligent systems like robots, visual understanding has gained vast interests in the past few decades. Most of the existing approaches need human labelled training data to train the learning models for visual understanding and in the most recent years, significant performance gain was obtained relying on unparalleled tremendous amount of human labelled training data. Under this circumstance, people are endowed with great burden to cost energy and time on the tedious data annotation for the traditional visual understanding approaches. To alleviate this problem, we propose to develop novel visual understanding algorithms which can learn informative visual patterns under minimal (none or very weak) supervision and thus facilitate higher-level intelligence of the visual understanding systems. Specifically, we focus on three subtopics, i.e., saliency detection, co-saliency detection, and weakly supervised learning based object detection, which can be used in both the image and video understanding systems. The experimental results have demonstrated the effectiveness of the proposed algorithms.
【Keywords】:
【Paper Link】 【Pages】:4050-4053
【Authors】: Shivani Agarwal
【Abstract】: In today's big data era, huge amounts of ranking and choice data are generated on a daily basis, and consequently, many powerful new computational tools for dealing with ranking and choice data have emerged in recent years. This paper highlights recent developments in two areas of ranking and choice modeling that cross traditional boundaries and are of multidisciplinary interest: ranking from pairwise comparisons, and automatic discovery of latent categories from choice survey data.
【Keywords】:
【Paper Link】 【Pages】:4054-4057
【Authors】: Haris Aziz
【Abstract】: Computational social choice is an exciting interdisciplinary field at the intersection of computer science and social choice theory. In this article, I discuss some current and new directions in the field. This is an accompanying paper of my IJCAI 2016 Early Career Spotlight invited talk.
【Keywords】:
【Paper Link】 【Pages】:4058-4061
【Authors】: Meghyn Bienvenu
【Abstract】: Ontology-mediated query answering (OMQA) is a new paradigm in data management that seeks to exploit the semantic knowledge expressed in ontologies to improve query answering over data. This paper briefly introduces OMQA and gives an overview of two recent lines of research.
【Keywords】:
【Paper Link】 【Pages】:4062-4065
【Authors】: Edith Elkind ; Martin Lackner ; Dominik Peters
【Abstract】: The goal of this short paper is to provide an overview of recent progress in understanding and exploiting useful properties of restricted preference domains, such as, e.g., the domains of single-peaked, single-crossing and 1-Euclidean preferences.
【Keywords】:
【Paper Link】 【Pages】:4066-4069
【Authors】: Matti Järvisalo
【Abstract】: This overview accompanies the author's IJCAI-16 Early Career Spotlight Talk, highlighting aspects of the author's research agenda with a strong focus on some of the author's recent research contributions.
【Keywords】:
【Paper Link】 【Pages】:4070-4073
【Authors】: Ece Kamar
【Abstract】: Hybrid intelligence systems combine machine and human intelligence to overcome the shortcomings of existing AI systems. This paper reviews recent research efforts towards developing hybrid systems focusing on reasoning methods for optimizing access to human intelligence and on gaining comprehensive understanding of humans as helpers of AI systems. It concludes by discussing short and long term research directions.
【Keywords】:
【Paper Link】 【Pages】:4074-4077
【Authors】: Mausam
【Abstract】: Open Information Extraction (Open IE) extracts textual tuples comprising relation phrases and argument phrases from within a sentence, without requiring a pre-specified relation vocabulary. In this paper we first describe a decade of our progress on building Open IE extractors, which results in our latest extractor, OpenIE4, which is computationally efficient, outputs n-ary and nested relations, and also outputs relations mediated by nouns in addition to verbs. We also identify several strengths of the Open IE paradigm, which enable it to be a useful intermediate structure for end tasks. We survey its use in both human-facing applications and downstream NLP tasks, including event schema induction, sentence similarity, text comprehension, learning word vector embeddings, and more.
【Keywords】:
【Paper Link】 【Pages】:4078-4081
【Authors】: Steven Schockaert ; Shoaib Jameel
【Abstract】: Formalizing and automating aspects of human plausible reasoning is an important challenge for the field of artificial intelligence. Practical advances, however, are hampered by the fact that most forms of plausible reasoning rely on background knowledge that is often not available in a structured form. In this paper, we first discuss how an important class of background knowledge can be induced from vector space representations that have been learned from (mostly) unstructured data. Subsequently, we advocate the use of qualitative abstractions of these vector spaces, as they are easier to obtain and manipulate, among others, while still supporting various forms of plausible reasoning.
【Keywords】:
【Paper Link】 【Pages】:4082-4085
【Authors】: Dafna Shahaf
【Abstract】: What can only humans do? What can humans do that machines cannot? These questions have long been tantalizing scientists and philosophers. Many areas, such as creativity and humor, are traditionally considered to be outside the reach of computers. We believe that these territories define an intriguing set of challenges for computer science. We present two approaches for tackling such challenges — an axiomatic one and a data-driven one — and demonstrate our ideas on two real-world applications: finding narratives in large textual corpora and identifying humorous cartoon captions.
【Keywords】:
【Paper Link】 【Pages】:4086-4089
【Authors】: Guy Van den Broeck
【Abstract】: First-order model counting recently emerged as a computational tool for high-level probabilistic reasoning. It is concerned with counting satisfying assignments to sentences in first-order logic and upgrades the successful propositional model counting approaches to probabilistic reasoning. We give an overview of model counting as it is applied in statistical relational learning, probabilistic programming, databases, and hybrid reasoning. A short tutorial illustrates the principles behind these solvers. Finally, we show that first-order counting is a fundamentally different problem from the propositional counting techniques that inspired it.
【Keywords】:
【Paper Link】 【Pages】:4090-4093
【Authors】: Pradeep Varakantham
【Abstract】: Rapid urbanization (more than 50% of worlds' population now resides in cities) coupled with the natural lack of coordination in usage of common resources (ex: bikes, ambulances, taxis, traffic personnel, attractions) has a detrimental effect on a wide variety of response (ex: waiting times, response time for emergency needs) and coverage metrics (ex: predictability of traffic/security patrols) in cities of today. Motivated by the need to improve response and coverage metrics in urban environments, my research group is focussed on building intelligent agent systems that make sequential decisions to continuously match available supply of resources to an uncertain demand for resources. Our broad approach to generating these sequential decision strategies is through a combination of data analytics (to obtain a model) and multi-stage optimization (planning/scheduling) under uncertainty (to solve the model). While we perform data analytics, our contributions are focussed on multi-stage optimization under uncertainty. We exploit key properties of urban environments, namely homogeneity and anonymity, limited influence of individual entities, abstraction and near decomposability to solve "multi-stage optimization under uncertainty" effectively and efficiently.
【Keywords】:
【Paper Link】 【Pages】:4094-4099
【Authors】: Yevgeniy Vorobeychik
【Abstract】: In recent years AI research has had an increasing role in models and algorithms for security problems.Game theoretic models of security, and Stackelberg security games in particular, have received special attention, in part because these models and associated tools have seen actual deployment in homeland security and sustainability applications. Stackelberg security games have two prototypical features: 1) a collection of potential assets which require protection, and 2) a sequential structure,where a defender first allocates protection resources, and the attacker then responds with an optimal attack.I see the latter feature as the major conceptual breakthrough,allowing very broad application of the idea beyond physical security settings.In particular, I describe three research problems which on the surface look nothing like prototypical security games: adversarial machine learning, privacy-preserving data sharing, and vaccine design.I describe how the second conceptual aspect of security games offers a natural modeling paradigm for these.This, in turn, has two important benefits: first, it offers a new perspective on these problems, and second, facilitates fundamental algorithmic contributions for these domains.
【Keywords】:
【Paper Link】 【Pages】:4100-4194
【Authors】: Mario Alviano ; Wolfgang Faber ; Martin Gebser
【Abstract】: In answer set programming, knowledge involving sets of objects collectively is naturally represented by aggregates, which are rewritten into simpler forms known as monotone aggregates by current implementations. However, there is a complexity gap between general and monotone aggregates. In this paper, this gap is filled by means of a polynomial, faithful, and modular translation function, which can introduce disjunction in rule heads. The translation function is now part of the recent version 4.5 of the grounder Gringo. This paper focuses on the key points of the translation function, and in particular on the mapping from non-convex sums to monotone sums.
【Keywords】:
【Paper Link】 【Pages】:4105-4109
【Authors】: Mario Alviano ; Nicola Leone
【Abstract】: Gelfond and Zhang recently proposed a new stable model semantics based on Vicious Circle Principle in order to improve the interpretation of logic programs with aggregates. A detailed complexity analysis of coherence testing and cautious reasoning under the new semantics highlighted similarities and differences versus mainstream stable model semantics for aggregates, which eventually led to the design of compilation techniques for implementing the new semantics on top of existing ASP solvers.
【Keywords】:
【Paper Link】 【Pages】:4110-4114
【Authors】: Angelos Angelidakis ; Georgios Chalkiadakis
【Abstract】: Tackling the decision-making problem faced by a prosumer (i.e., a producer that is simultaneously a consumer) when selling and buying energy in the emerging smart electricity grid, is of utmost importance for the economic profitability of such a business entity. In this work, we model, for the first time, this problem as a factored Markov Decision Process. By so doing, we are able to rep-resent the problem compactly, and provide an ex-act optimal solution via dynamic programming — notwithstanding its large size. Our model success-fully captures the main aspects of the business decisions of a prosumer corresponding to a community microgrid of any size. Moreover, it includes appropriate sub-models for prosumer production and consumption prediction. Experimental simulations verify the effectiveness of our approach; and show that our exact value iteration solution matches that of a state-of-the-art method for stochastic planning in very large environments, while outperforming it in terms of computation time.
【Keywords】:
【Paper Link】 【Pages】:4115-4119
【Authors】: Vaishak Belle ; Guy Van den Broeck ; Andrea Passerini
【Abstract】: In recent years, there has been considerable progress on fast randomized algorithms that approximate probabilistic inference with tight tolerance and confidence guarantees. The idea here is to formulate inference as a counting task over an annotated propositional theory, called weighted model counting (WMC), which can be partitioned into smaller tasks using universal hashing. An inherent limitation of this approach, however, is that it only admits the inference of discrete probability distributions. In this work, we consider the problem of approximating inference tasks for a probability distribution defined over discrete and continuous random variables. Building on a notion called weighted model integration, which is a strict generalization of WMC and is based on annotating Boolean and arithmetic constraints, we show how probabilistic inference in hybrid domains can be put within reach of hashing-based WMC solvers. Empirical evaluations demonstrate the applicability and promise of the proposal.
【Keywords】:
【Paper Link】 【Pages】:4120-4124
【Authors】: Alina Beygelzimer ; Satyen Kale ; Haipeng Luo
【Abstract】: We study online boosting, the task of converting any weak online learner into a strong online learner. Based on a novel and natural definition of weak online learnability, we develop two online boosting algorithms. The first algorithm is an online version of boost-by-majority. By proving a matching lower bound, we show that this algorithm is essentially optimal in terms of the number of weak learners and the sample complexity needed to achieve a specified accuracy. The second algorithm is adaptive and parameter-free, albeit not optimal.
【Keywords】:
【Paper Link】 【Pages】:4125-4129
【Authors】: Nigel Bosch ; Sidney K. D'Mello ; Ryan S. Baker ; Jaclyn Ocumpaugh ; Valerie Shute ; Matthew Ventura ; Lubin Wang ; Weinan Zhao
【Abstract】: Affect detection is a key component of intelligent educational interfaces that can respond to the affective states of students. We use computer vision, learning analytics, and machine learning to detect students' affect in the real-world environment of a school computer lab that contained as many as thirty students at a time. Students moved around, gestured, and talked to each other, making the task quite difficult. Despite these challenges, we were moderately successful at detecting boredom, confusion, delight, frustration, and engaged concentration in a manner that generalized across students, time, and demographics. Our model was applicable 98% of the time despite operating on noisy real-world data.
【Keywords】:
【Paper Link】 【Pages】:4130-4134
【Authors】: Abdeslam Boularias ; Felix Duvallet ; Jean Oh ; Anthony Stentz
【Abstract】: We consider the problem of robots following natural language commands through previously unknown outdoor environments. A robot receives commands in natural language, such as Navigate around the building to the car left of the fire hydrant and near the tree. The robot needs first to classify its surrounding objects into categories, using images obtained from its sensors. The result of this classification is a map of the environment, where each object is given a list of semantic labels, such as tree or car, with varying degrees of confidence. Then, the robot needs to ground the nouns in the command, i.e., mapping each noun in the command into a physical object in the environment. The robot needs also to ground a specified navigation mode, such as navigate quickly or navigate covertly, as a cost map. In this work, we show how to ground nouns and navigation modes by learning from examples demonstrated by humans.
【Keywords】:
【Paper Link】 【Pages】:4135-4139
【Authors】: Martin C. Cooper ; Achref El Mouelhi ; Cyril Terrioux ; Bruno Zanuttini
【Abstract】: A binary CSP instance satisfying the broken-triangle property (BTP) can be solved in polynomial time. Unfortunately, in practice, few instances satisfy the BTP. We show that a local version of the BTP allows the merging of domain values in binary CSPs, thus providing a novel polynomial-time reduction operation. Experimental trials on benchmark instances demonstrate a significant decrease in instance size for certain classes of problems. We show that BTP-merging can be generalised to instances of arbitrary arity. A directional version of the general-arity BTP then allows us to extend the BTP tractable class previously defined only for binary CSP.
【Keywords】:
【Paper Link】 【Pages】:4140-4144
【Authors】: Toby O. Davies ; Adrian R. Pearce ; Peter J. Stuckey ; Nir Lipovetzky
【Abstract】: Operator-counting is a recently developed framework for analysing and integrating many state-of-the-art heuristics for planning using Linear Programming. In cost-optimal planning only the objective value of these heuristics is traditionally used to guide the search. However the primal solution, i.e. the operator counts, contains useful information. We exploit this information using a SAT-based approach which given an operator-count, either finds a valid plan; or generates a generalized landmark constraint violated by that count. We show that these generalized landmarks can be used to encode the perfect heuristic, h*, as a Mixed Integer Program. Our most interesting experimental result is that finding or refuting a sequence for an operator-count is most often empirically efficient, enabling a novel and promising approach to planning based on Logic-Based Benders Decomposition (LBBD).
【Keywords】:
【Paper Link】 【Pages】:4145-4149
【Authors】: Benoit Desouter ; Marko van Dooren ; Tom Schrijvers ; Alexander Vandenbroucke
【Abstract】: The logic programming language Prolog uses a resource-efficient SLD resolution strategy for query answering. Yet, its propensity for nontermination seriously detracts from the language's declarative nature. This problem is remedied by tabling, a modified execution strategy that allows a larger class of programs to terminate. Unfortunately, few Prolog systems provide tabling, because the documented implementation techniques are complex, low-level and require a prohibitive engineering effort. To enable more widespread adoption, this paper presents a novel implementation of tabling for Prolog that is both high-level and compact. It comes in the form of a Prolog library that weighs in at under 600 lines of code, is based on delimited control and delivers reasonable performance.
【Keywords】:
【Paper Link】 【Pages】:4150-4154
【Authors】: Vijay D'Silva ; Caterina Urban
【Abstract】: One can prove that a program satisfies a correctness property in different ways. The deductive approach uses logic and is automated using decision procedures and proof assistants. The automata-theoretic approach reduces questions about programs to algorithmic questions about automata. In the abstract interpretation approach, programs and their properties are expressed in terms of fixed points in lattices and reasoning uses fixed point approximation techniques. We describe a research programme to establish precise, mathematical correspondences between these approaches and to develop new analyzers using these results. The theoretical tools we use are the theorems of Büchi that relate automata and logic and a construction of Lindenbaum and Tarski for generating lattices from logics. This research has lead to improvements in existing tools and we anticipate further theoretical and practical consequences.
【Keywords】:
【Paper Link】 【Pages】:4155-4159
【Authors】: Guillem Francès ; Hector Geffner
【Abstract】: Most of the key computational ideas in classical planning assume a simple planning language where action preconditions and goals are conjunctions of propositional atoms. This is to facilitate the definition and computation of heuristics for guiding the search for plans. In this work, however, we show that this modeling choice hides important structural information, resulting in poorer heuristics and weaker planning performance. To address this, we show how relaxed plan heuristics can be lifted to a variable-free first-order planning language, Functional STRIPS, where atomic formulas can involve arbitrary terms. The key idea is to regard the set of atoms that are reachable in a propositional layer of the relaxed planning graph as encoding a set of logical first-order interpretations. A preconditionor goal formula is then regarded as reachable in a propositional layer, potentially adding new atoms to the next layer, when the set of atoms in the layer makes the formula satisfiable according to the rules of first-order logic. While this satisfiability test and the resulting heuristics turn out to be intractable, we show how a meaningful polynomial approximation can be obtained by formulating the satisfiability problem as a CSP and applying constraint propagation techniques. Experiments illustrating the computational value of planning with more expressive languages are also reported.
【Keywords】:
【Paper Link】 【Pages】:4160-4164
【Authors】: Peter Gregory ; Stephen Cresswell
【Abstract】: We present a new domain model acquisition algorithm, LOP, that induces static predicates by using a combination of the generalised output from LOCM2 and a set of optimal plans as input to the learning system. We observe that static predicates can be seen as restrictions on the valid groundings of actions. Without the static predicates restricting possible groundings, the domains induced by LOCM2 produce plans that are typically shorter than the true optimal solutions. LOP works by finding a set of minimal static predicates for each operator that preserves the length of the optimal plan.
【Keywords】:
【Paper Link】 【Pages】:4165-4169
【Authors】: Chinmay Hegde ; Piotr Indyk ; Ludwig Schmidt
【Abstract】: We introduce a framework for sparsity structures defined via graphs. Our approach is flexible and generalizes several previously studied sparsity models. Moreover, we provide efficient projection algorithms for our sparsity model that run in nearly-linear time. In the context of sparse recovery, our framework achieves an information-theoretically optimal sample complexity for a wide range of parameters. We complement our theoretical analysis with experiments showing that our algorithms also improve on prior work in practice.
【Keywords】:
【Paper Link】 【Pages】:4170-4174
【Authors】: Joshua Hernandez ; Konstantine Tsotsos ; Stefano Soatto
【Abstract】: We analyze the observability of 3-D position and orientation from the fusion of visual and inertial sensors. The model contains unknown parameters, such as sensor biases, and so the problem is usually cast as a mixed filtering/identification problem, with the resulting observability analysis providing necessary conditions for convergence to a unique point estimate. Most models treat sensor bias rates as noise, independent of other states, including biases themselves, an assumption that is violated in practice. We show that, when this assumption is lifted, the resulting model is not observable, and therefore existing analyses cannot be used to conclude that the set of states that are indistinguishable from the measurements is a singleton. We recast the analysis as one of sensitivity: Rather than attempting to prove that the set of indistinguishable trajectories is a singleton, we derive bounds on its volume, as a function of characteristics of the sensor and other sufficient excitation conditions. This provides an explicit characterization of the indistinguishable set that can be used for analysis and validation purposes.
【Keywords】:
【Paper Link】 【Pages】:4175-4179
【Authors】: John N. Hooker
【Abstract】: Projection can be seem as a unifying concept that underlies inference in logic and consistency maintenance in constraint programming. This perspective allows one to import projection methods into both areas, resulting in deeper insight as well as faster solution methods. We show that inference in propositional logic can be achieved by Benders decomposition, an optimization method based on projection. In constraint programming, viewing consistency maintenance as projection suggests a new but natural concept of consistency that is achieved by projection onto a subset of variables. We show how to solve this combinatorial projection problem for some global constraints frequently used in constraint programming. The resulting projections are useful when propagated through decision diagrams rather than the traditional domain store.
【Keywords】:
【Paper Link】 【Pages】:4180-4189
【Authors】: Nan Jiang ; Alex Kulesza ; Satinder P. Singh ; Richard L. Lewis
【Abstract】: Because planning with a long horizon (i.e., looking far into the future) is computationally expensive, it is common in practice to save time by using reduced horizons. This is usually understood to come at the expense of computing suboptimal plans, which is the case when the planning model is exact. However, when the planning model is estimated from data, as is frequently true in the real world, the policy found using a shorter planning horizon can actually be better than a policy learned with the true horizon. In this paper we provide a precise explanation for this phenomenon based on principles of learning theory. We show formally that the planning horizon is a complexity control parameter for the class of policies available to the planning algorithm, having an intuitive, monotonic relationship with a simple measure of complexity. We prove a planning loss bound predicting that shorter planning horizons can reduce overfitting and improve test performance, and we confirm these predictions empirically.
【Keywords】:
【Paper Link】 【Pages】:4185-4189
【Authors】: Philip Kilby ; Tommaso Urli
【Abstract】: We present an original approach to compute efficient mid-term fleet configurations at the request of a Queensland-based long-haul trucking carrier. Our approach considers one year's worth of demand data, and employs a constraint programming (CP) model and an adaptive large neighbourhood search (LNS) scheme to solve the underlying multi-day multi-commodity split delivery capacitated vehicle routing problem. This paper is an adaptation of the Best Application Paper at CP'15, published in the Constraints journal with the same title.
【Keywords】:
【Paper Link】 【Pages】:4190-4194
【Authors】: Peter Kontschieder ; Madalina Fiterau ; Antonio Criminisi ; Samuel Rota Bulò
【Abstract】: We present a novel approach to enrich classification trees with the representation learning ability of deep (neural) networks within an end-to-end trainable architecture. We combine these two worlds via a stochastic and differentiable decision tree model, which steers the formation of latent representations within the hidden layers of a deep network. The proposed model differs from conventional deep networks in that a decision forest provides the final predictions and it differs from conventional decision forests by introducing a principled, joint and global optimization of split and leaf node parameters. Our approach compares favourably to other state-of-the-art deep models on a large-scale image classification task like ImageNet.
【Keywords】:
【Paper Link】 【Pages】:4195-4199
【Authors】: Bo Liu ; Ji Liu ; Mohammad Ghavamzadeh ; Sridhar Mahadevan ; Marek Petrik
【Abstract】: In this paper, we describe proximal gradient temporal difference learning, which provides a principled way for designing and analyzing true stochastic gradient temporal difference learning algorithms. We show how gradient TD (GTD) reinforcement learning methods can be formally derived, not with respect to their original objective functions as previously attempted, but rather with respect to primal-dual saddle-point objective functions. We also conduct a saddle-point error analysis to obtain finite-sample bounds on their performance. Previous analyses of this class of algorithms use stochastic approximation techniques to prove asymptotic convergence, and no finite-sample analysis had been attempted. An accelerated algorithm is also proposed, namely GTD2-MP, which use proximal mirror maps to yield acceleration. The results of our theoretical analysis imply that the GTD family of algorithms are comparable and may indeed be preferred over existing least squares TD methods for off-policy learning, due to their linear complexity. We provide experimental results showing the improved performance of our accelerated gradient TD methods.
【Keywords】:
【Paper Link】 【Pages】:4200-4204
【Authors】: Morten Mossige ; Arnaud Gotlieb ; Hein Meling
【Abstract】: Designing industrial robot systems for welding, painting, and assembly, is challenging because they must perform with high precision, speed, and endurance. ABB Robotics has specialized in building highly reliable and safe robotized paint systems using an integrated process control system. However, current validation practices are mainly limited to manual test scenarios, which makes it difficult to exercise important aspects of a paint robot system, such as the need to coordinate the timing of paint activation with the robot motion control. To address these challenges, we have developed and deployed a cost-effective, automated test generation technique aimed at validating the timing behavior of the process control system. The approach is based on a constraint optimization model written in Prolog. This model has been integrated into an automated continuous integration environment, allowing the model to be solved on demand prior to test execution, which allows us to obtain the most optimal and diverse set of test scenarios for the current system configuration.
【Keywords】:
【Paper Link】 【Pages】:4205-4209
【Authors】: Andrew Reynolds ; Jasmin Christian Blanchette
【Abstract】: Datatypes and codata types are useful to represent finite and potentially infinite objects. We describe a decision procedure to reason about such types. The procedure has been integrated into CVC4, a modern SMT (satisfiability modulo theories) solver, which can be used both as a constraint solver and as an automatic theorem prover. An evaluation based on formalizations developed in the Isabelle proof assistant shows the potential of the procedure.
【Keywords】:
【Paper Link】 【Pages】:4210-4212
【Authors】: Tim Roughgarden ; Inbal Talgam-Cohen
【Abstract】: Computational complexity has already had plenty to say about the computation of economic equilibria [Fischer et al., 2006; Chen et al., 2009b; 2009a; Daskalakis et al., 2009; Papadimitriou and Wilkens, 2011]. However, understanding when equilibria are guaranteed to exist is a central theme in economic theory, seemingly unrelated to computation. In this note we survey our main results from [Roughgarden and Talgam-Cohen, 2015], which show that the existence of equilibria in markets is inextricably connected to the computational complexity of related optimization problems, such as revenue or welfare maximization. We demonstrate how this relationship implies, under suitable complexity assumptions, a host of impossibility results. We also suggest a complexity-theoretic explanation for the lack of useful extensions of the Walrasian equilibrium concept: such extensions seem to require the invention of novel polynomial-time algorithms for welfare maximization.
【Keywords】:
【Paper Link】 【Pages】:4213-4217
【Authors】: Wen Sun ; J. Andrew Bagnell
【Abstract】: We establish connections from optimizing Bellman Residual and Temporal Difference Loss to worst-case long-term predictive error. In the online learning framework, learning takes place over a sequence of trials with the goal of predicting a future discounted sum of rewards. Our first analysis shows that, together with a stability assumption, any no-regret online learning algorithm that minimizes Bellman error ensures small prediction error. Our second analysis shows that applying the family of online mirror descent algorithms on temporal difference loss also ensures small prediction error. No statistical assumptions are made on the sequence of observations, which could be non-Markovian or even adversarial. Our approach thus establishes a broad new family of provably sound algorithms and provides a generalization of previous worst-case results for minimizing predictive error. We investigate the potential advantages of some of this family both theoretically and empirically on benchmark problems.
【Keywords】:
【Paper Link】 【Pages】:4218-4222
【Authors】: Elaine Wah ; Mason Wright ; Michael P. Wellman
【Abstract】: We investigate the effects of market making on market performance, focusing on allocative efficiency as well as gains from trade accrued by background traders. We employ empirical simulation-based methods to evaluate heuristic strategies for market makers as well as background investors in a variety of complex trading environments. Our market model incorporates private and common valuation elements, with dynamic fundamental value and asymmetric information. In this context, we compare the surplus achieved by background traders in strategic equilibrium, with and without a market maker. Our findings indicate that the presence of the market maker strongly tends to increase total welfare across a variety of environments. Market-maker profit may or may not exceed the welfare gain, thus the effect on background-investor surplus is ambiguous. We find that market making tends to benefit investors in relatively thin markets, and situations where background traders are impatient, due to limited trading opportunities. Introducing additional market makers increases these benefits, as competition drives market makers to provide liquidity at lower price spreads.
【Keywords】:
【Paper Link】 【Pages】:4223-4227
【Authors】: Yi Yang ; Shimei Pan ; Yangqiu Song ; Jie Lu ; Mercan Topkara
【Abstract】: Topic modeling has become a ubiquitous topic analysis tool for text exploration. Most of the existing works on topic modeling focus on fitting topic models to input data. They however ignore an important usability issue that is closely related to the end user experience: stability. In this study, we investigate the stability problem in topic modeling. We first report on the experiments conducted to quantify the severity of the problem. We then propose a new learning framework to mitigate the problem by explicitly incorporating topic stability constraints in model training. We also perform user study to demonstrate the advantages of the proposed method.
【Keywords】:
【Paper Link】 【Pages】:4228-4233
【Authors】: Edward Zulkoski ; Vijay Ganesh ; Krzysztof Czarnecki
【Abstract】: We present a method and an associated system, called MathCheck, that embeds the functionality of a computer algebra system (CAS) within the inner loop of a conflict-driven clause-learning SAT solver. SAT+CAS systems, a la MathCheck, can be used as an assistant by mathematicians to either counterexample or finitely verify open universal conjectures on any mathematical topic (e.g., graph and number theory, algebra, geometry, etc.) supported by the underlying CAS system. Such a SAT+CAS system combines the efficient search routines of modern SAT solvers, with the expressive power of CAS, thus complementing both. The key insight behind the power of the SAT+CAS combination is that the CAS system can help cut down the search-space of the SAT solver, by providing learned clauses that encode theory-specific lemmas, as it searches for a counterexample to the input conjecture. We demonstrate the efficacy of our approach on a long-standing open conjecture regarding matchings of hypercubes.
【Keywords】:
【Paper Link】 【Pages】:4234-4235
【Authors】: C. Anantaram ; Sunil Kumar Kopparapu ; Chiragkumar Patel ; Aditya Mittal
【Abstract】: General-purpose speech engines are trained on large corpus. However, studies and experiments have shown that when such engines are used to recognize spoken sentences in specific domains they may not produce accurate ASR output. Further, the accent and the environmental conditions in which the speaker speaks a sentence may induce the speech engine to recognize certain words/ sets of words inaccurately. Thus, the speech engine's output may need to be repaired for a domain before any further natural language processing is carried out. We present an artificial development (Art-Dev) based mechanism for such a repair. Our approach considers an erroneous ASR output sentence as a biological cell and repairs it through evolution and development of the inaccurate genes in the cell (sentence) with respect to the genes in the domain. Once the genotypes are identified, we grow the genotypes into phenotypes to fill the missing gaps or erroneous words with appropriate domain concepts. We demonstrate our approach on the output of standard ASR engines such as Google Now and show how it improves the accuracy.
【Keywords】:
【Paper Link】 【Pages】:4236-4237
【Authors】: Roman Barták ; Michal Koutný ; David Obdrzálek
【Abstract】: There exist solutions for tracking of objects in 3D space involving hi-tech cameras and powerful computer systems capable of tracking many objects in large dynamic space simultaneously in real time. On the other hand, there are situations where such functionality is not necessary and the conditions may be specified in more detail, which makes the task significantly easier. This paper shows the possibility to track a single object using low-cost cameras on an ordinary laptop in a small-scale and mostly static environment. This solution is useful for standalone tracking in mobile robotics and particularly in the debugging phases, where the user needs to judge the robot movement system independently on what the robot claims.
【Keywords】:
【Paper Link】 【Pages】:4238-4239
【Authors】: Noam Brown ; Tuomas Sandholm
【Abstract】: Imperfect-information games, where players have private information, pose a unique challenge in artificial intelligence. In recent years, Heads-Up No-Limit Texas Hold'em poker, a popular version of poker, has emerged as the primary benchmark for evaluating game-solving algorithms for imperfect-information games. We demonstrate a winning agent from the 2016 Annual Computer Poker Competition, Baby Tartanian8.
【Keywords】:
【Paper Link】 【Pages】:4240-4241
【Authors】: Wanyun Cui ; Yanghua Xiao ; Wei Wang
【Abstract】: Question answering (QA) has become a popular way for humans to access billion-scale knowledge bases. QA systems over knowledge bases produce accurate and concise answers. The key of QA over knowledge bases is to map the question to a certain substructure in the knowledge base. To do this, KBQA (Question Answering over Knowledge Bases) uses a new kind of question representation: templates, learned from a million scale QA corpora. For example, for questions about a city's population, KBQA learns templates such as What's the population of $city?, How many people are there in $city?. It learns overall 1171303 templates for 4690 relations. Based on these templates, KBQA effectively and efficiently supports binary factoid questions or complex questions.
【Keywords】:
【Paper Link】 【Pages】:4242-4243
【Authors】: Mauro Dragoni ; Célia da Costa Pereira ; Andrea G. B. Tettamanzi ; Serena Villata
【Abstract】: The extraction of the relevant and debated opinions from online social media and commercial websites is an emerging task in the opinion mining research field. Its growing relevance is mainly due to the impact of exploiting such techniques in different application domains from social science analysis to personal advertising.In this demo, we present our opinion summary application built on top of an argumentation framework, a standard AI framework whose value is to exchange, communicate and resolve possibly conflicting viewpoints in distributed scenarios. We show how our application is able to extract relevant and debated opinions from a set of documents containing user-generated content from online commercial websites.
【Keywords】:
【Paper Link】 【Pages】:4244-4245
【Authors】: Michael Freed ; Brian Burns ; Aaron Heller ; Daniel Sanchez ; Sharon Beaumont-Bowman
【Abstract】: For millions of people with swallowing disorders, preventing potentially deadly aspiration pneumonia requires following prescribed safe eating strategies. But adherence is poor, and caregivers' ability to encourage adherence is limited by the onerous and socially aversive need to monitoring another's eating. We have developed an early prototype for an intelligent assistant that monitors adherence and provides feedback to the patient, and tested monitoring precision with healthy subjects for one strategy called a chin tuck. Results indicate that adaptations of current generation machine vision and personal assistant technologies can effectively monitor chin tuck adherence, and suggest feasibility of a more general assistant that encourages adherence to a range of safe eating strategies.
【Keywords】:
【Paper Link】 【Pages】:4246-4247
【Authors】: Matthew Johnson ; Katja Hofmann ; Tim Hutton ; David Bignell
【Abstract】: We present Project Malmo - an AI experimentation platform built on top of the popular computer game Minecraft, and designed to support fundamental research in artificial intelligence. As the AI research community pushes for artificial general intelligence (AGI), experimentation platforms are needed that support the development of flexible agents that learn to solve diverse tasks in complex environments. Minecraft is an ideal foundation for such a platform, as it exposes agents to complex 3D worlds, coupled with infinitely varied game-play. Project Malmo provides a sophisticated abstraction layer on top of Minecraft that supports a wide range of experimentation scenarios, ranging from navigation and survival to collaboration and problem solving tasks. In this demo we present the Malmo platform and its capabilities. The platform is publicly released as open source software at IJCAI, to support openness and collaboration in AI research.
【Keywords】:
【Paper Link】 【Pages】:4248-4249
【Authors】: James Kirk ; Aaron Mininger ; John E. Laird
【Abstract】: We will demonstrate a tabletop robotic agent that learns new tasks through interactive natural language instruction. The tasks to be demonstrated are simple puzzles and games, such as Tower of Hanoi, Eight Puzzle, Tic-Tac-Toe, Three Men's Morris, and the Frog and Toads puzzle. We will include a live, interactive simulation of a mobile robot that learns new tasks using the same system.
【Keywords】:
【Paper Link】 【Pages】:4250-4251
【Authors】: Sanmukh R. Kuppannagari ; Rajgopal Kannan ; Charalampos Chelmis ; Viktor K. Prasanna
【Abstract】: Demand Response (DR) allows utilities to curtail electricity consumption during peak demand periods. Real time automated DR can offer utilities a scalable solution for fine grained control of curtailment over small intervals for the duration of the entire DR event. In this work, we demonstrate a system for a real time automated Dynamic DR (D2R). Our system has already been integrated with the electrical infrastructure of the University of Southern California, which offers a unique environment to study the impact of automated DR in a complex social and cultural environment including 170 buildings in a city-within-a-city scenario. Our large scale information processing system coupled with accurate forecasting models for sparse data and fast polynomial time optimization algorithms for curtailment maximization provide the ability to adapt and respond to changing curtailment requirements in near real-time. Our D2 R algorithms automatically and dynamically select customers for load curtailment to guarantee the achievement of a curtailment target over a given DR interval.
【Keywords】:
【Paper Link】 【Pages】:4252-4253
【Authors】: Domenico Lembo ; Daniele Pantaleone ; Valerio Santarelli ; Domenico Fabio Savo
【Abstract】: We demonstrate Eddy, a new tool for designing ontologies specified in the Graphol language. Graphol is completely visual and fully captures OWL 2. Thus Eddy is the first ontology editor that allows to create OWL 2 ontologies by only using simple graphical editing features.
【Keywords】:
【Paper Link】 【Pages】:4254-4255
【Authors】: Chao-Chun Liang ; Kuang-Yi Hsu ; Chien-Tsung Huang ; Chung-Min Li ; Shen-Yu Miao ; Keh-Yih Su
【Abstract】: This demonstration presents a tag-based statistical English math word problem (MWP) solver with understanding, reasoning, and explanation. It analyzes the text and transforms both body and question parts into their tag-based logic forms, and then performs inference on them. The proposed tag (e.g., Agent, Verb, etc.) provides the flexibility for annotating an extracted math quantity with its associated syntactic and semantic information, which can be used to identify the desired operand and filter out irrelevant quantities (so that the answer can be obtained precisely). Since the physical meaning of each quantity is explicitly represented with those tags and used in the inference process, the proposed approach could explain how the answer is obtained in a human comprehensible way.
【Keywords】:
【Paper Link】 【Pages】:4256-4257
【Authors】: Yu Lu ; Gim Guan Chua ; Huayu Wu ; Clement Shi Qi Ong
【Abstract】: The fast advancement in sensor data acquisition and communication technology greatly facilitates the collection of data from taxis, and thus enables analyzing the citywide taxi service system. In this paper, we present a novel and practical system for taxi service monitoring, analytics and visualization. By utilizing both of the buffered streaming and the large-size historical taxi data, the system focuses on wait time estimation (for both passengers and taxi drivers), citywide taxi pickup/dropoff hotspots, as well as the taxi trip distributions. The three-dimensional (3D) visualization is designed for users to access the analytics results and understand the characteristics of the taxi service.
【Keywords】:
【Paper Link】 【Pages】:4258-4259
【Authors】: Andrea Marrella ; Massimo Mecella ; Sebastian Sardiña
【Abstract】: In this paper, we introduce an adaptive Process Management System implementation that combines business process execution monitoring, unanticipated exception detection and automated resolution strategies leveraging on well-established formalisms developed for reasoning about actions in Artificial Intelligence, including the Situation Calculus, IndiGolog and classical planning. Such formalisms provide a natural framework for the formal specification of explicit mechanisms to model world changes and responding to anomalous situations, exceptions, exogenous events in an automated way during process execution.
【Keywords】:
【Paper Link】 【Pages】:4260-4261
【Authors】: J. Pablo Munoz ; Bing Li ; Xuejian Rong ; Jizhong Xiao ; Yingli Tian ; Aries Arditi
【Abstract】: Research in Artificial Intelligence, Robotics and Computer Vision has recently made great strides in improving indoor localization. Publicly available technology now allows for indoor localization with very small margins of error. In this demo, we show a system that uses state-of the-art technology to as- sist visually impaired people navigate indoors. Our system takes advantage of spatial representations from CAD files, or floor plan images, to extract valuable information that later can be used to im- prove navigation and human-computer interaction. Using depth information, our system is capable of detecting obstacles and guiding the user to avoid them.
【Keywords】:
【Paper Link】 【Pages】:4262-4263
【Authors】: Jacobo Rouces ; Gerard de Melo ; Katja Hose
【Abstract】: An increasing number of structured knowledge bases have become available on the Web, enabling many new forms of analyses and applications. However, the fact that the data is being published by different parties with different vocabularies and ontologies means that there is a high level of heterogeneity and no common schema. This paper presents Klint, a web-based system that automatically creates mappings to transform knowledge as provided by the sources into data that conforms to a large unified schema. The user can review and edit the mappings with a streamlined interface. In this way, Klint allows for human-level accuracy with minimum human effort.
【Keywords】:
【Paper Link】 【Pages】:4264-4265
【Authors】: Muhammad Rizwan Saeed ; Charalampos Chelmis ; Viktor K. Prasanna
【Abstract】: The combination of data, semantics, and the Web has led to an ever growing and increasingly complex body of semantic data. Accessing such structured data requires learning formal query languages, such as SPARQL, which poses significant difficulties for non-expert users. Many existing interfaces for querying Ontologies are based on approaches that rely on predefined templates and require expensive customization. To avoid the pitfalls of existing approaches, while at the same time retaining the ability to capture users' complex information needs, we have developed a simple keyword-based search interface to the Semantic Web. In this demonstration, we will present ASQFor, a systematic framework for automated SPARQL query formulation and execution over RDF repository using simple concept-based search primitives. Allowing end-users to express simple queries based on a list of "key-value" pairs that are then translated on-the-fly into SPARQL queries is a hard problem. In this demonstration, we will discuss the challenges that we have addressed to bring ASQFor to real practice, and also the difficult problems that remain to be solved in future work. During our demonstration, we will show how ASQFor can be used for decision support as well as an intelligent Q/A System.
【Keywords】:
【Paper Link】 【Pages】:4266-4267
【Authors】: François Schwarzentruber
【Abstract】: We describe a tool for generating Euler diagrams from a set of region connection calculus formulas. The generation is based on a variant of local search capturing default reasoning for improving aesthetic appearance of Euler diagrams. We also describe an optimization for diagrams to be interactive: the user can modify the diagram with the mouse while formulas are still satisfied. We also discuss how such a tool may propose new relevant formulas to add to the specification using an approximation algorithm based on the satisfiability of Horn clauses.
【Keywords】:
【Paper Link】 【Pages】:4268-4269
【Authors】: Shirin Sohrabi ; Octavian Udrea ; Anton V. Riabov ; Oktie Hassanzadeh
【Abstract】: We present LTS++, an interactive development environment for planning-based hypothesis generation in applications with unreliable observations.
【Keywords】:
【Paper Link】 【Pages】:4270-4271
【Authors】: Srikanth Tamilselvam ; Biplav Srivastava ; Vishalaksh Aggarwal
【Abstract】: We demonstrate a family of novel, standards-based, online applications for promoting tourist events with minimal operational impact using AI methods. The capabilities span collection of events using standards and crowd, their analysis to promote discovery by future tourists as well as manage impact by city managers, and its dissemination.The solution offers benefits to citizens, travelers, city managers and businesses and can be rolled out to cities around the world.
【Keywords】:
【Paper Link】 【Pages】:4272-4273
【Authors】: Feiyi Tang ; Jia Zhu ; Yang Cao ; Sanli Ma ; Yulong Chen ; Jing He ; Changqin Huang ; Gansen Zhao ; Yong Tang
【Abstract】: Widely adoption of GPS-enabled devices generates massive trajectory data every minute. The trajectory data can generate meaningful traffic patterns. In this demo, we present a system called PARecommender, which predicts traffic conditions and provides route recommendation based on generated traffic patterns. We first introduce the technical details of PARecommender, and then show several real cases that how PARecommender works.
【Keywords】:
【Paper Link】 【Pages】:4274-4275
【Authors】: Donghan Wang ; Madalina Fiterau ; Artur Dubrawski
【Abstract】: Analysis, pattern discovery, and decision support can all benefit greatly from informative and interpretable visualizations, especially of high-dimensional data. Informative Projection Ensemble (IPE) methodology has proven effective in finding interpretable renderings of high-dimensional data that reveal hidden low-dimensional structures in data if such structures exist. In this demonstration, we present a powerful analysis tool that uses IPE methodology in support of fundamental machine learning tasks: regression, classification, and clustering. Our tool is an interactive web application operating on 2D and 3D projections of data automatically selected by IPE algorithms as informative for the user-specified data and task.It also provides RESTful APIs enabling remote users to seamlessly integrate our service with other tools and to easily extend its functionality. We show in examples how it can discover hidden interpretable structures embedded in high-dimensional data.
【Keywords】:
【Paper Link】 【Pages】:4276-4277
【Authors】: Shiqi Zhang ; Dongcai Lu ; Xiaoping Chen ; Peter Stone
【Abstract】: In recent years, many different types of intelligent mobile robots have been developed in research and industrial labs. Although there are significant differences in both hardware and software over these robots, many of them share a common set of AI capabilities, e.g., planning, learning, vision and natural language processing. At the same time, almost all of them are equipped with traditional robotic capabilities such as mapping, localization, and navigation. However, to date it has been difficult to compare and contrast their capabilities in any controlled way. The main goal of the Robot Scavenger Hunt is to provide a standardized framework that includes a set of standardized tasks for evaluating the AI and robotic capabilities of medium-sized intelligent mobile robots. Compared to existing benchmarks, e.g., RoboCup@Home, Robot Scavenger Hunt aims at evaluations in larger spaces (multi-floor buildings vs. rooms) over longer periods of time (hours vs.minutes) while interacting with real human residents.
【Keywords】: