25. AAAI 2011:San Francisco, California, USA

Proceedings of the Twenty-Fifth AAAI Conference on Artificial Intelligence, AAAI 2011, San Francisco, California, USA, August 7-11, 2011. AAAI Press 【DBLP Link

Paper Num: 319 || Session Num: 18

1. Extensible Automated Constraint Modelling.

Paper Link】 【Pages】:

【Authors】: Ozgur Akgun ; Ian Miguel ; Christopher Jefferson ; Alan M. Frisch ; Brahim Hnich

【Abstract】: In constraint solving, a critical bottleneck is the formulation of aneffective constraint model of an input problem. The Conjure system describedin this paper, a substantial step forward over prototype versions of Conjurepreviously reported, makes a valuable contribution to the automation ofconstraint modelling by automatically producing constraint models from theirspecifications in the abstract constraint specification language Essence. Aset of rules is used to refine an abstract specification into a concreteconstraint model. We demonstrate that this set of rules is readily extensibleto increase the space of possible constraint models Conjure can produce. Ourempirical results confirm that Conjure can reproduce successfully the kernelsof the constraint models of 32 benchmark problems found in the literature.

【Keywords】:

Paper Link】 【Pages】:

【Authors】: David Burkett ; David Leo Wright Hall ; Dan Klein

【Abstract】: Informed search algorithms such as A use heuristics to focus exploration on states with low total path cost. To the extent that heuristics underestimate forward costs, a wider cost radius of suboptimal states will be explored. For many weighted graphs, however, a small distance in terms of cost may encompass a large fraction of the unweighted graph. We present a new informed search algorithm, Iterative Monotonically Bounded A (IMBA), which first proves that no optimal paths exist in a bounded cut of the graph before considering larger cuts. We prove that IMBA has the same optimality and completeness guarantees as A and, in a non-uniform pathfinding application, we empirically demonstrate substantial speed improvements over classic A.

【Keywords】:

3. On the Complexity of BDDs for State Space Search: A Case Study in Connect Four.

Paper Link】 【Pages】:

【Authors】: Stefan Edelkamp ; Peter Kissmann

【Abstract】: Symbolic search using BDDs usually saves huge amounts of memory, while in some domains its savings are moderate at best. It is an open problem to determine if BDDs work well for a certain domain. Motivated by finding evidences for BDD growths for state space search, in this paper we are concerned with symbolic search in the domain of Connect Four. We prove that there is a variable ordering for which the set of all possible states – when continuing after a terminal state has been reached – can be represented by polynomial sized BDDs, whereas the termination criterion leads to an exponential number of nodes in the BDD given any variable ordering.

【Keywords】:

4. The Compressed Differential Heuristic.

Paper Link】 【Pages】:

【Authors】: Meir Goldenberg ; Nathan R. Sturtevant ; Ariel Felner ; Jonathan Schaeffer

【Abstract】: The differential heuristic (DH) is an effective memory-based heuristic for explicit state spaces. In this paper we aim to improve its performance and memory usage. We introduce a compression method for DHs which stores only a portion of the original uncompressed DH, while preserving enough information to enable efficient search. Compressed DHs (CDH) are flexible and can be tuned to fit any size of memory, even smaller than the size of the state space. Furthermore, CDHs can be built without the need to create and store the entire uncompressed DH. Experimental results across different domains show that, for a given amount of memory, a CDH significantly outperforms an uncompressed DH.

【Keywords】:

Paper Link】 【Pages】:

【Authors】: Matthew Hatem ; Ethan Burns ; Wheeler Ruml

【Abstract】: The memory requirements of basic best-first heuristic search algorithms like A make them infeasible for solving large problems. External disk storage is cheap and plentiful com- pared to the cost of internal RAM. Unfortunately, state-of- the-art external memory search algorithms either rely on brute-force search techniques, such as breadth-first search, or they rely on all node values falling in a narrow range of in- tegers, and thus perform poorly on real-world domains with real-valued costs. We present a new general-purpose algo- rithm, PEDAL, that uses external memory and parallelism to perform a best-first heuristic search capable of solving large problems with real costs. We show theoretically that PEDAL is I/O efficient and empirically that it is both better on a stan- dard unit-cost benchmark, surpassing internal IDA on the 15-puzzle, and gives far superior performance on problems with real costs.

【Keywords】:

Paper Link】 【Pages】:

【Authors】: Federico Heras ; António Morgado ; João Marques-Silva

【Abstract】: Several MaxSAT algorithms based on iterative SAT solving have been proposed in recent years. These algorithms are in general the most efficient for real-world applications. Existing data indicates that, among MaxSAT algorithms based on iterative SAT solving, the most efficient ones are core-guided, i.e. algorithms which guide the search by iteratively computing unsatisfiable subformulas (or cores). For weighted MaxSAT, core-guided algorithms exhibit a number of important drawbacks, including a possibly exponential number of iterations and the use of a large number of auxiliary variables. This paper develops two new algorithms for (weighted) MaxSAT that address these two drawbacks. The first MaxSAT algorithm implements core-guided iterative SAT solving with binary search. The second algorithm extends the first one by exploiting disjoint cores. The empirical evaluation shows that core-guided binary search is competitive with current MaxSAT solvers.

【Keywords】:

7. Optimal Packing of High-Precision Rectangles.

Paper Link】 【Pages】:

【Authors】: Eric Huang ; Richard E. Korf

【Abstract】: The rectangle-packing problem consists of finding an enclosing rectangle of smallest area that can contain a given set of rectangles without overlap. Our new benchmark includes rectangles of successively higher precision, challenging the previous state-of-the-art, which enumerates all locations for placing rectangles, as well as all bounding box widths and heights up to the optimal box. We instead limit the rectangles’ coordinates and bounding box dimensions to the set of subset sums of the rectangles’ dimensions. We also dynamically prune values by learning from infeasible subtrees and constrain the problem by replacing rectangles and empty space with larger rectangles. Compared to the previous state-of-the-art, we test 4,500 times fewer bounding boxes on the high-precision benchmark and solve N =9 over two orders of magnitude faster. We also present all optimal solutions up to N =15, which requires 39 bits of precision to solve. Finally, on the open problem of whether or not one can pack a particular infinite series of rectangles into the unit square, we pack the first 50,000 such rectangles witha greedy heuristic and conjecture that the entire infinite series can fit..

【Keywords】:

8. A General Nogood-Learning Framework for Pseudo-Boolean Multi-Valued SAT.

Paper Link】 【Pages】:

【Authors】: Siddhartha Jain ; Ashish Sabharwal ; Meinolf Sellmann

【Abstract】: We formulate a general framework for pseudo-Boolean multi-valued nogood-learning, generalizing conflict analysis performed by modern SAT solvers and its recent extension for disjunctions of multi-valued variables. This framework can handle more general constraints as well as different domain representations, such as interval domains which are commonly used for bounds consistency in constraint programming (CP), and even set variables. Our empirical evaluation shows that our solver, built upon this framework, works robustly across a number of challenging domains.

【Keywords】:

9. Pushing the Power of Stochastic Greedy Ordering Schemes for Inference in Graphical Models.

Paper Link】 【Pages】:

【Authors】: Kalev Kask ; Andrew Gelfand ; Lars Otten ; Rina Dechter

【Abstract】: We study iterative randomized greedy algorithms for generating (elimination) orderings with small induced width and state space size — two parameters known to bound the complexity of inference in graphical models. We propose and implement the Iterative Greedy Variable Ordering (IGVO) algorithm, a new variant within this algorithm class. An empirical evaluation using different ranking functions and conditions of randomness, demonstrates that IGVO finds significantly better orderings than standard greedy ordering implementations when evaluated within an anytime framework. Additional order of magnitude improvements are demonstrated on a multi-core system, thus further expanding the set of solvable graphical models. The experiments also confirm the superiority of the MinFill heuristic within the iterative scheme.

【Keywords】:

10. A Comparison of Lex Bounds for Multiset Variables in Constraint Programming.

Paper Link】 【Pages】:

【Authors】: Yat Chiu Law ; Jimmy Ho-Man Lee ; May Hiu-Chun Woo ; Toby Walsh

【Abstract】: Set and multiset variables in constraint programming have typically been represented using subset bounds. However, this is a weak representation that neglects potentially useful information about a set such as its cardinality. For set variables, the length-lex (LL) representation successfully provides information about the length (cardinality) and position in the lexicographic ordering. For multiset variables, where elements can be repeated, we consider richer representations that take into account additional information. We study eight different representations in which we maintain bounds according to one of the eight different orderings: length-(co)lex (LL/LC), variety-(co)lex (VL/VC), length-variety-(co)lex (LVL/LVC), and variety-length-(co)lex (VLL/VLC) orderings. These representations integrate together information about the cardinality, variety (number of distinct elements in the multiset), and position in some total ordering. Theoretical and empirical comparisons of expressiveness and compactness of the eight representations suggest that length-variety-(co)lex (LVL/LVC) and variety-length-(co)lex (VLL/VLC) usually give tighter bounds after constraint propagation. We implement the eight representations and evaluate them against the subset bounds representation with cardinality and variety reasoning. Results demonstrate that they offer significantly better pruning and runtime.

【Keywords】:

11. Distributed Constraint Optimization Under Stochastic Uncertainty.

Paper Link】 【Pages】:

【Authors】: Thomas Léauté ; Boi Faltings

【Abstract】: In many real-life optimization problems involving multiple agents, the rewards are not necessarily known exactly in advance, but rather depend on sources of exogenous uncertainty. For instance, delivery companies might have to coordinate to choose who should serve which foreseen customer, under uncertainty in the locations of the customers. The framework of Distributed Constraint Optimization under Stochastic Uncertainty was proposed to model such problems; in this paper, we generalize this formalism by introducing the concept of evaluation functions that model various optimization criteria. We take the example of three such evaluation functions, expectation , consensus , and robustness , and we adapt and generalize two previous algorithms accordingly. Our experimental results on a class of Vehicle Routing Problems show that incomplete algorithms are not only cheaper than complete ones (in terms of simulated time , Non-Concurrent Constraint Checks , and information exchange) , but they are also often able to find the optimal solution. We also show that exchanging more information about the dependencies of their respective cost functions on the sources of uncertainty can help the agents discover higher-quality solutions.

【Keywords】:

12. Planning in Domains with Cost Function Dependent Actions.

Paper Link】 【Pages】:

【Authors】: Mike Phillips ; Maxim Likhachev

【Abstract】: In a number of graph search-based planning problems, the value of the cost function that is being minimized also affects the set of possible actions at some or all the states in the graph. For example, in path planning for a robot with a limited battery power, a common cost function is energy consumption, whereas the level of remaining energy affects the navigational capabilities of the robot. Similarly, in path planning for a robot navigating dynamic environments, a total traversal time is a common cost function whereas the timestep affects whether a particular transition is valid. In such planning problems, the cost function typically becomes one of the state variables thereby increasing the dimensionality of the planning problem, and consequently the size of the graph that represents the problem. In this paper, we show how to avoid this increase in the dimensionality for the planning problems whenever the availability of the actions is monotonically non-increasing with the increase in the cost function. We present three variants of A* search for dealing with such planning problems: a provably optimal version, a suboptimal version that scales to larger problems while maintaining a bound on suboptimality, and finally a version that relaxes our assumption on the relationship between the cost function and action space. Our experimental analysis on several domains shows that the presented algorithms achieve up to several orders of magnitude speed up over the alternative approaches to planning.

【Keywords】:

13. Euclidean Heuristic Optimization.

Paper Link】 【Pages】:

【Authors】: D. Chris Rayner ; Michael H. Bowling ; Nathan R. Sturtevant

【Abstract】: We pose the problem of constructing good search heuristics as an optimization problem: minimizing the loss between the true distances and the heuristic estimates subject to admissibility and consistency constraints. For a well-motivated choice of loss function, we show performing this optimization is tractable. In fact, it corresponds to a recently proposed method for dimensionality reduction. We prove this optimization is guaranteed to produce admissible and consistent heuristics, generalizes and gives insight into differential heuristics, and show experimentally that it produces strong heuristics on problems from three distinct search domains.

【Keywords】:

14. Succinct Set-Encoding for State-Space Search.

Paper Link】 【Pages】:

【Authors】: Tim Schmidt ; Rong Zhou

【Abstract】: We introduce the level-ordered edge sequence (LOES), a suc- cinct encoding for state-sets based on prefix-trees. For use in state-space search, we give algorithms for member testing and element hashing with runtime dependent only on state- size, as well as space and memory efficient construction of and iteration over such sets. Finally we compare LOES to binary decision diagrams (BDDs) and explicitly packed set- representation over a range of IPC planning problems. Our results show LOES produces succinct set-encodings for a wider range of planning problems than both BDDs and ex- plicit state representation, increasing the number of problems that can be solved cost-optimally.

【Keywords】:

15. Limits of Preprocessing.

Paper Link】 【Pages】:

【Authors】: Stefan Szeider

【Abstract】: We present a first theoretical analysis of the power of polynomial-time preprocessing for important combinatorial problems from various areas in AI. We consider problems from Constraint Satisfaction, Global Constraints, Satisfiability, Nonmonotonic and Bayesian Reasoning. We show that, subject to a complexity theoretic assumption, none of the considered problems can be reduced by polynomial-time preprocessing to a problem kernel whose size is polynomial in a structural problem parameter of the input, such as induced width or backdoor size. Our results provide a firm theoretical boundary for the performance of polynomial-time preprocessing algorithms for the considered problems.

【Keywords】:

16. Inner Regions and Interval Linearizations for Global Optimization.

Paper Link】 【Pages】:

【Authors】: Gilles Trombettoni ; Ignacio Araya ; Bertrand Neveu ; Gilles Chabert

【Abstract】: Researchers from interval analysis and constraint (logic) programming communities have studied intervals for their ability to manage infinite solution sets of numerical constraint systems. In particular, inner regions represent subsets of the search space in which all points are solutions. Our main contribution is the use of recent and new inner region extraction algorithms in the upper bounding phase of constrained global optimization. Convexification is a major key for efficiently lower bounding the objective function. We have adapted the convex interval taylorization proposed by Lin and Stadherr for producing a reliable outer and inner polyhedral approximation of the solution set and  a linearization of the objective function. Other original ingredients are part of our optimizer, including an efficient interval constraint propagation algorithm exploiting monotonicity of functions. We end up with a new framework for reliable continuous constrained global optimization. Our interval B&B is implemented in the interval-based explorer Ibex and extends this free C++ library. Our strategy outperforms the best reliable global optimizers.

【Keywords】:

17. Anytime Nonparametric A.

Paper Link】 【Pages】:

【Authors】: Jur van den Berg ; Rajat Shah ; Arthur Huang ; Kenneth Y. Goldberg

【Abstract】: Anytime variants of Dijkstra's and A shortest path algorithms quickly produce a suboptimal solution and then improve it over time. For example, ARA introduces a weighting value "epsilon" to rapidly find an initial suboptimal path and then reduces "epsilon" to improve path quality over time. In ARA, "epsilon" is based on a linear trajectory with ad-hoc parameters chosen by each user. We propose a new Anytime A algorithm, Anytime Nonparametric A (ANA), that does not require ad-hoc parameters, and adaptively reduces varepsilon to expand the most promising node per iteration, adapting the greediness of the search as path quality improves. We prove that each node expanded by ANA provides an upper bound on the suboptimality of the current-best solution. We evaluate the performance of ANA with experiments in the domains of robot motion planning, gridworld planning, and multiple sequence alignment. The results suggest that ANA is as efficient as ARA and in most cases: (1) ANA finds an initial solution faster, (2) ANA spends less time between solution improvements, (3) ANA decreases the suboptimality bound of the current-best solution more gradually, and (4) ANA finds the optimal solution faster. ANA* is freely available from Maxim Likhachev's Search-based Planning Library (SBPL).

【Keywords】:

18. Solving Difficult CSPs with Relational Neighborhood Inverse Consistency.

Paper Link】 【Pages】:

【Authors】: Robert J. Woodward ; Shant Karakashian ; Berthe Y. Choueiry ; Christian Bessiere

【Abstract】: Freuder and Elfe (1996) introduced Neighborhood Inverse Consistency (NIC) as a strong local consistency property for binary CSPs. While enforcing NIC can significantly filter the variables domains, the proposed algorithm is too costly to be used on dense graphs or for lookahead during search. In this paper, we introduce and characterize Relational Neighborhood Inverse Consistency (RNIC) as a local consistency property that operates on the dual graph of a non-binary CSP. We describe and characterize a practical algorithm for enforcing it. We argue that defining RNIC on the dual graph unveils unsuspected opportunities to reduce the computational cost of our algorithm and increase its filtering effectiveness. We show how to achieve those effects by modifying the topology of the dual graph, yielding new variations the RNIC property. We also introduce an adaptive strategy to automatically select the appropriate property to enforce given the connectivity of the dual graph. We integrate the resulting techniques as full lookahead strategies in a backtrack search procedure for solving CSPs, and demonstrate the effectiveness of our approach for solving known difficult benchmark problems.

【Keywords】:

Paper Link】 【Pages】:

【Authors】: Peter Yap ; Neil Burch ; Robert C. Holte ; Jonathan Schaeffer

【Abstract】: We present three new ideas for grid-based path-planning algorithms that improve the search speed and quality of the paths found. First, we introduce a new type of database, the Local Distance Database (LDDB), that contains distances between boundary points of a local neighborhood. Second, an LDDB based algorithm is introduced, called Block A, that calculates the optimal path between start and goal locations given the local distances stored in the LDDB. Third, our experimental results for any-angle path planning in a wide variety of test domains, including real game maps, show that Block A is faster than both A and the previously best grid-based any-angle search algorithm, Theta.

【Keywords】:

Knowledge-Based Information Systems 6

20. Simulated Annealing Based Influence Maximization in Social Networks.

Paper Link】 【Pages】:

【Authors】: Qingye Jiang ; Guojie Song ; Gao Cong ; Yu Wang ; Wenjun Si ; Kunqing Xie

【Abstract】: The problem of influence maximization, i.e., mining top-k influential nodes from a social network such that the spread of influence in the network is maximized, is NP-hard. Most of the existing algorithms for the prob- lem are based on greedy algorithm. Although greedy algorithm can achieve a good approximation, it is computational expensive. In this paper, we propose a totally different approach based on Simulated Annealing(SA) for the influence maximization problem. This is the first SA based algorithm for the problem. Additionally, we propose two heuristic methods to accelerate the con- vergence process of SA, and a new method of comput- ing influence to speed up the proposed algorithm. Experimental results on four real networks show that the proposed algorithms run faster than the state-of-the-art greedy algorithm by 2-3 orders of magnitude while being able to improve the accuracy of greedy algorithm.

【Keywords】:

21. Tracking User-Preference Varying Speed in Collaborative Filtering.

Paper Link】 【Pages】:

【Authors】: Ruijiang Li ; Bin Li ; Cheng Jin ; Xiangyang Xue ; Xingquan Zhu

【Abstract】: In real-world recommender systems, some users are easily influenced by new products and whereas others are unwilling to change their minds. So the preference varying speeds for users are different. Based on this observation, we propose a dynamic nonlinear matrix factorization model for collaborative filtering, aimed to improve the rating prediction performance as well as track the preference varying speeds for different users. We assume that user-preference changes smoothly over time, and the preference varying speeds for users are different. These two assumptions are incorporated into the proposed model as prior knowledge on user feature vectors, which can be learned efficiently by MAP estimation. The experimental results show that our method not only achieves state-of-the-art performance in the rating prediction task, but also provides an effective way to track user-preference varying speed.

【Keywords】:

22. CosTriage: A Cost-Aware Triage Algorithm for Bug Reporting Systems.

Paper Link】 【Pages】:

【Authors】: Jin-Woo Park ; Mu-Woong Lee ; Jinhan Kim ; Seung-won Hwang ; Sunghun Kim

【Abstract】: "Who can fix this bug?" is an important question in bug triage to "accurately" assign developers to bug reports. To address this question, recent research treats it as a optimizing recommendation accuracy problem and proposes a solution that is essentially an instance of content-based recommendation (CBR). However, CBR is well-known to cause over-specialization, recommending only the types of bugs that each developer has solved before. This problem is critical in practice, as some experienced developers could be overloaded, and this would slow the bug fixing process. In this paper, we take two directions to address this problem: First,we reformulate the problem as an optimization problem of both accuracy and cost. Second, we adopt a content-boosted collaborative filtering (CBCF), combining an existing CBR with a collaborative filtering recommender (CF), which enhances the recommendationquality of either approach alone. However, unlike general recommendation scenarios, bug fix history is extremely sparse. Due to the nature of bug fixes, one bug is fixed by only one developer, which makes it challenging to pursue the above two directions. To address this challenge, we develop a topic-model to reduce the sparseness and enhance the quality of CBCF. Our experimental evaluation shows that our solution reduces the cost efficiently by 30% without seriously compromising accuracy.

【Keywords】:

23. Relational Blocking for Causal Discovery.

Paper Link】 【Pages】:

【Authors】: Matthew J. Rattigan ; Marc E. Maier ; David Jensen

【Abstract】: Blocking is a technique commonly used in manual statistical analysis to account for confounding variables. However, blocking is not currently used in automated learning algorithms. These algorithms rely solely on statistical conditioning as an operator to identify conditional independence. In this work, we present relational blocking as a new operator that can be used for learning the structure of causal models. We describe how blocking is enabled by relational data sets, where blocks are determined by the links in the network. By blocking on entities rather than conditioning on variables, relational blocking can account for both measured and unobserved variables. We explain the mechanism of these methods using graphical models and the semantics of d-separation. Finally, we demonstrate the effectiveness of relational blocking for use in causal discovery by showing how blocking can be used in the causal analysis of two real-world social media systems.

【Keywords】:

24. Deriving a Web-Scale Common Sense Fact Database.

Paper Link】 【Pages】:

【Authors】: Niket Tandon ; Gerard de Melo ; Gerhard Weikum

【Abstract】: The fact that birds have feathers and ice is cold seems trivially true. Yet, most machine-readable sources of knowledge either lack such common sense facts entirely or have only limited coverage. Prior work on automated knowledge base construction has largely focused on relations between named entities and on taxonomic knowledge, while disregarding common sense properties. In this paper, we show how to gather large amounts of common sense facts from Web n-gram data, using seeds from the ConceptNet collection. Our novel contributions include scalable methods for tapping onto Web-scale data and a new scoring model to determine which patterns and facts are most reliable. The experimental results show that this approach extends ConceptNet by many orders of magnitude at comparable levels of precision.

【Keywords】:

25. Social Recommendation Using Low-Rank Semidefinite Program.

Paper Link】 【Pages】:

【Authors】: Jianke Zhu ; Hao Ma ; Chun Chen ; Jiajun Bu

【Abstract】: The most critical challenge for the recommendation system is to achieve the high prediction quality on the large scale sparse data contributed by the users. In this paper, we present a novel approach to the social recommendation problem, which takes the advantage of the graph Laplacian regularization to capture the underlying social relationship among the users. Differently from the previous approaches, that are based on the conventional gradient descent optimization, we formulate the presented graph Laplacian regularized social recommendation problem into a low-rank semidefinite program, which is able to be efficiently solved by the quasi-Newton algorithm. We have conducted the empirical evaluation on a large scale dataset of high sparsity, the promising experimental results show that our method is very effective and efficient for the social recommendation task.

【Keywords】:

Knowledge Representation and Reasoning 20

26. A Semantical Account of Progression in the Presence of Uncertainty.

Paper Link】 【Pages】:

【Authors】: Vaishak Belle ; Gerhard Lakemeyer

【Abstract】: Building on a general theory of action by Reiter and his colleagues, Bacchus et al. give an account for formalizing degrees of belief and noisy actions in the situation calculus. Unfortunately, there is no clear solution to the projection problem for the formalism. And, while the model has epistemic features, it is not obvious what the agent's knowledge base should look like. Also, reasoning about uncertainty essentially resorts to second-order logic. In recent work, Gabaldon and Lakemeyer remedy these shortcomings somewhat, but here too the utility seems to be restricted to queries (with action operators) about the initial theory. In this paper, we propose a fresh amalgamation of a modal fragment of the situation calculus and uncertainty, where the idea will be to update the initial knowledge base, containing both ordinary and (certain kinds of) probabilistic beliefs, when noisy actions are performed. We show that the new semantics has the right properties, and study a special case where updating probabilistic beliefs is computable. Our ideas are closely related to the Lin and Reiter notion of progression.

【Keywords】:

27. Adding Default Attributes to EL++.

Paper Link】 【Pages】:

【Authors】: Piero A. Bonatti ; Marco Faella ; Luigi Sauro

【Abstract】: The research on low-complexity nonmonotonic description logics recently identified a fragment of EL with bottom, supporting defeasible inheritance with overriding, where reasoning can be carried out in polynomial time. We contribute to that framework by supporting more axiom schemata and all the concept constructors of EL++ without increasing asymptotic complexity. Moreover, we show that all the syntactic restrictions we adopt are necessary by proving several coNP-hardness results.

【Keywords】:

28. Learning from Spatial Overlap.

Paper Link】 【Pages】:

【Authors】: Michael H. Coen ; M. Hidayath Ansari ; Nathanael Fillmore

【Abstract】: This paper explores a new measure of similarity between point sets in arbitrary metric spaces. The measure is based on the spatial overlap of the “shapes” and “densities” of these point sets. It is applicable in any domain where point sets are a natural representation for data. Specifically, we show examples of its use in natural language processing, object recognition in images and point set classification. We provide a geometric interpretation of this measure and show that it is well-motivated, intuitive, parameter-free, and straightforward to use. We further demonstrate that it is computationally tractable and applicable to both supervised and unsupervised learning problems.

【Keywords】:

29. Higher-Order Description Logics for Domain Metamodeling.

Paper Link】 【Pages】:

【Authors】: Giuseppe De Giacomo ; Maurizio Lenzerini ; Riccardo Rosati

【Abstract】: We investigate an extension of Description Logics (DL) with higher-order capabilities, based on Henkin-style semantics. Our study starts from the observation that the various possibilities of adding higher-order con- structs to a DL form a spectrum of increasing expres- sive power, including domain metamodeling, i.e., using concepts and roles as predicate arguments. We argue that higher-order features of this type are sufficiently rich and powerful for the modeling requirements aris- ing in many relevant situations, and therefore we carry out an investigation of the computational complexity of satisfiability and conjunctive query answering in DLs extended with such higher-order features. In particular, we show that adding domain metamodeling capabilities to SHIQ (the core of OWL 2) has no impact on the complexity of the various reasoning tasks. This is also true for DL-LiteR (the core of OWL 2 QL) under suit- able restrictions on the queries.

【Keywords】:

30. Spectrum-Based Sequential Diagnosis.

Paper Link】 【Pages】:

【Authors】: Alberto González-Sanchez ; Rui Abreu ; Hans-Gerhard Groß ; Arjan J. C. van Gemund

【Abstract】: We present a spectrum-based, sequential software debugging approach coined Sequoia, that greedily selects tests out of a suite of tests to narrow down the set of diagnostic candidates with a minimum number of tests. Sequoia handles multiple faults, that can be intermittent, at polynomial time and space complexity, due to a novel, approximate diagnostic entropy estimation approach, which considers the subset of diagnoses that cover almost all Bayesian posterior probability mass. Synthetic experiments show that Sequoia achieves much better diagnostic uncertainty reduction compared to random test sequencing.Real programs, taken from the Software Infrastructure Repository, confirm Sequoia's better performance, with a test reduction up to 80% compared to random test sequences.

【Keywords】:

31. A Closer Look at the Probabilistic Description Logic Prob-EL.

Paper Link】 【Pages】:

【Authors】: Víctor Gutiérrez-Basulto ; Jean Christoph Jung ; Carsten Lutz ; Lutz Schröder

【Abstract】: We study probabilistic variants of the description logic EL. For the case where probabilities apply only to concepts, we provide a careful analysis of the borderline between tractability and ExpTime-completeness. One outcome is that any probability value except zero and one leads to intractability in the presence of general TBoxes, while this is not the case for classical TBoxes. For the case where probabilities can also be applied to roles, we show PSpace-completeness. This result is (positively) surprising as the best previously known upper bound was 2-ExpTime and there were reasons to believe in completeness for this class.

【Keywords】:

32. Trajectory Regression on Road Networks.

Paper Link】 【Pages】:

【Authors】: Tsuyoshi Idé ; Masashi Sugiyama

【Abstract】: This paper addresses the task of trajectory cost prediction, a new learning task for trajectories. The goal of this task is to predict the cost for an arbitrary (possibly unknown) trajectory, based on a set of previous trajectory-cost pairs. A typical example of this task is travel-time prediction on road networks. The main technical challenge here is to infer the costs of trajectories including links with no or little passage history. To tackle this, we introduce a weight propagation mechanism over the links, and show that the problem can be reduced to a simple form of kernel ridge regression. We also show that this new formulation leads us to a unifying view, where a natural choice of the kernel is suggested to an existing kernel-based alternative.

【Keywords】:

33. An Algebraic Prolog for Reasoning about Possible Worlds.

Paper Link】 【Pages】:

【Authors】: Angelika Kimmig ; Guy Van den Broeck ; Luc De Raedt

【Abstract】: We introduce aProbLog, a generalization of the probabilistic logic programming language ProbLog. An aProbLog program consists of a set of definite clauses and a set of algebraic facts; each such fact is labeled with an element of a semiring. A wide variety of labels is possible, ranging from probability values to reals (representing costs or utilities), polynomials, Boolean functions or data structures. The semiring is then used to calculate labels of possible worlds and of queries. We formally define the semantics of aProbLog and study the aProbLog inference problem, which is concerned with computing the label of a query. Two conditions are introduced that allow one to simplify the inference problem, resulting in four different algorithms and settings. Representative basic problems for each of these four settings are: is there a possible world where a query is true (SAT), how many such possible worlds are there (#SAT), what is the probability of a query being true (PROB), and what is the most likely world where the query is true (MPE). We further illustrate these settings with a number of tasks requiring more complex semirings.

【Keywords】:

34. Two-Dimensional Description Logics for Context-Based Semantic Interoperability.

Paper Link】 【Pages】:

【Authors】: Szymon Klarman ; Víctor Gutiérrez-Basulto

【Abstract】: Description Logics (DLs) provide a clear and broadly accepted paradigm for modeling and reasoning about terminological knowledge. However, it has been often noted, that although DLs are well-suited for representing a single, global viewpoint on an application domain, they offer no formal grounding for dealing with knowledge pertaining to multiple heterogeneous viewpoints — a scenario ever more often approached in practical applications, e.g. concerned with reasoning over distributed knowledge sources on the Semantic Web. In this paper, we study a natural extension of DLs, in the style of two-dimensional modal logics, which supports declarative modeling of viewpoints as contexts, in the sense of McCarthy, and their semantic interoperability. The formalism is based on two-dimensional semantics, where one dimension represents a usual object domain and the other a (possibly infinite) domain of viewpoints, addressed by additional modal operators and a metalanguage, on the syntactic level. We systematically introduce a number of expressive fragments of the proposed logic, study their computational complexity and connections to related formalisms.

【Keywords】:

35. Conjunctive Query Inseparability of OWL 2 QL TBoxes.

Paper Link】 【Pages】:

【Authors】: Boris Konev ; Roman Kontchakov ; Michel Ludwig ; Thomas Schneider ; Frank Wolter ; Michael Zakharyaschev

【Abstract】: The OWL 2 profile OWL 2 QL, based on the DL-Lite family of description logics, is emerging as a major language for developing new ontologies and approximating the existing ones. Its main application is ontology-based data access, where ontologies are used to provide background knowledge for answering queries over data. We investigate the corresponding notion of query inseparability (or equivalence) for OWL 2 QL ontologies and show that deciding query inseparability is PSPACE-hard and in EXPTIME. We give polynomial time (incomplete) algorithms and demonstrate by experiments that they can be used for practical module extraction.

【Keywords】:

36. A Modular Consistency Proof for DOLCE.

Paper Link】 【Pages】:

【Authors】: Oliver Kutz ; Till Mossakowski

【Abstract】: We propose a novel technique for proving the consistency of large, complex and heterogeneous theories for which ‘standard’ automated reasoning methods are considered insufficient. In particular, we exemplify the applicability of the method by establishing the consistency of the foundational ontology DOLCE, a large, first-order ontology. The approach we advocate constructs a global model for a theory, in our case DOLCE, built from smaller models of subtheories together with amalgamability properties between such models. The proof proceeds by (i) hand-crafting a so-called architectural specification of DOLCE which reflects the way models of the theory can be built, (ii) an automated verification of the amalgamability conditions, and (iii) a (partially automated) series of relative consistency proofs.

【Keywords】:

37. Causal Theories of Actions Revisited.

Paper Link】 【Pages】:

【Authors】: Fangzhen Lin ; Mikhail Soutchanski

【Abstract】: It has been argued that causal rules are necessary for representing both implicit side-effects of actions and action qualifications, and there have been a number different approaches for representing causal rules in the area of formal theoriesof actions. These different approaches in general agree on rules without cycles. However, they differ on causal rules with mutual cyclic dependencies, both in terms of how these rules are supposed to be represented and their semantics. In this paper we show that by adding one more minimization to Lin's circumscriptive causal theory in the situation calculus, we can have a uniform representation of causal rules including those with cyclic dependencies. We also demonstrate that sometimes causal rules can be compiled into logically equivalent successor state axioms even in the presence of cyclical dependencies between fluents.

【Keywords】:

38. Revisiting Semantics for Epistemic Extensions of Description Logics.

Paper Link】 【Pages】:

【Authors】: Anees Mehdi ; Sebastian Rudolph

【Abstract】: Epistemic extensions of description logics (DLs) have been introduced several years ago in order to enhance expressivity and querying capabilities of these logics by knowledge base introspection. We argue that unintended effects occur when imposing the traditionally employed semantics on the very expressive DLs that underly the OWL 1 and OWL 2 standards. Consequently, we suggest a revised semantics that behaves more intuitively in these cases and coincides with the traditional semantics of less expressive DLs. Moreover, we introduce a way of answering epistemic queries to OWL knowledge bases by a reduction to standard OWL reasoning. We provide an implementation of our approach and present first evaluation results.

【Keywords】:

39. Transportability of Causal and Statistical Relations: A Formal Approach.

Paper Link】 【Pages】:

【Authors】: Judea Pearl ; Elias Bareinboim

【Abstract】: We address the problem of transferring information learned from experiments to a different environment, in which only passive observations can be collected. We introduce a formal representation called "selection diagrams" for expressing knowledge about differences and commonalities between environments and, using this representation, we derive procedures for deciding whether effects in the target environment can be inferred from experiments conducted elsewhere. When the answer is affirmative, the procedures identify the set of experiments and observations that need be conducted to license the transport. We further discuss how transportability analysis can guide the transfer of knowledge in non-experimental learning to minimize re-measurement cost and improve prediction power.

【Keywords】:

40. How to Calibrate the Scores of Biased Reviewers by Quadratic Programming.

Paper Link】 【Pages】:

【Authors】: Magnus Roos ; Jörg Rothe ; Björn Scheuermann

【Abstract】: Peer reviewing is the key ingredient of evaluating the quality of scientific work. Based on the review scores assigned by the individual reviewers to the submissions, program committees of conferences and journal editors decide which papers to accept for publication and which to reject. However, some reviewers may be more rigorous than others, they may be biased one way or the other, and they often have highly subjective preferences over the papers they review. Moreover, each reviewer usually has only a very local view, as he or she evaluates only a small fraction of the submissions. Despite all these shortcomings, the review scores obtained need to be aggregrated in order to globally rank all submissions and to make the acceptance/rejection decision. A common method is to simply take the average of each submission's review scores, possibly weighted by the reviewers' confidence levels. Unfortunately, the global ranking thus produced often suffers a certain unfairness, as the reviewers' biases and limitations are not taken into account. We propose a method for calibrating the scores of reviewers that are potentially biased and blindfolded by having only partial information. Our method uses a maximum likelihood estimator, which estimates both the bias of each individual reviewer and the unknown "ideal" score of each submission. This yields a quadratic program whose solution transforms the individual review scores into calibrated, globally comparable scores. We argue why our method results in a fairer and more reasonable global ranking than simply taking the average of scores. To show its usefulness, we test our method empirically using real-world data.

【Keywords】:

41. Preferred Explanations: Theory and Generation via Planning.

Paper Link】 【Pages】:

【Authors】: Shirin Sohrabi ; Jorge A. Baier ; Sheila A. McIlraith

【Abstract】: In this paper we examine the general problem of generating preferred explanations for observed behavior with respect to a model of the behavior of a dynamical system. This problem arises in a diversity of applications including diagnosis of dynamical systems and activity recognition. We provide a logical characterization of the notion of an explanation. To generate explanations we identify and exploit a correspondence between explanation generation and planning. The determination of good explanations requires additional domain-specific knowledge which we represent as preferences over explanations. The nature of explanations requires us to formulate preferences in a somewhat retrodictive fashion by utilizing Past Linear Temporal Logic. We propose methods for exploiting these somewhat unique preferences effectively within state-of-the-art planners and illustrate the feasibility of generating (preferred) explanations via planning.

【Keywords】:

42. Language Splitting and Relevance-Based Belief Change in Horn Logic.

Paper Link】 【Pages】:

【Authors】: Maonian Wu ; Dongmo Zhang ; Mingyi Zhang

【Abstract】: This paper presents a framework for relevance-based belief change in propositional Horn logic. We firstly establish a parallel interpolation theorem for Horn logic and show that Parikh's Finest Splitting Theorem holds with Horn formulae. By reformulating Parikh's relevance criterion in the setting of Horn belief change, we construct a relevance-based partial meet Horn contraction operator and provide a representation theorem for the operator. Interestingly, we find that this contraction operator can be fully characterised by Delgrande and Wassermann's postulates for partial meet Horn contraction as well as Parikh's relevance postulate without requiring any change on the postulates, which is qualitatively different from the case in classical propositional logic.

【Keywords】:

43. Integrating Rules and Description Logics by Circumscription.

Paper Link】 【Pages】:

【Authors】: Qian Yang ; Jia-Huai You ; Zhiyong Feng

【Abstract】: We present a new approach to characterizing the semantics for the integration of rules and first-order logic in general, and description logics in particular, based on a circumscription characterization of answer set programming, introduced earlier by Lin and Zhou. We show that both Rosati's semantics based on NM-models and Lukasiewicz's answer set semantics can be characterized by circumscription, and the difference between the two can be seen as a matter of circumscription policies. This approach leads to a number of new insights. First, we rebut a criticism on Lukasiewicz's semantics for its inability to reason for negative consequences. Second, our approach leads to a spectrum of possible semantics based on different circumscription policies, and shows a clear picture of how they are related. Finally, we show that the idea of this paper can be applied to first-order general stable models.

【Keywords】:

44. Bounded Forgetting.

Paper Link】 【Pages】:

【Authors】: Yi Zhou ; Yan Zhang

【Abstract】: The result of forgetting some predicates in a first-order sentence may not exist in the sense that it might not be captured by any first-order sentences. This, indeed, severely restricts the usage of forgetting in applications. To address this issue, we propose a notion called $k$-forgetting, also called bounded forgetting in general, for any fixed number $k$. We present several equivalent characterizations of bounded forgetting and show that the result of bounded forgetting, on one hand, can always be captured by a single first-order sentence, and on the other hand, preserves the information that we are concerned with.

【Keywords】:

45. Progression Semantics for Disjunctive Logic Programs.

Paper Link】 【Pages】:

【Authors】: Yi Zhou ; Yan Zhang

【Abstract】: In this paper, we extend the progression semantics for first-order disjunctive logic programs and show that it coincides with the stable model semantics. Based on it, we further show how disjunctive answer set programming is related to Satisfiability Modulo Theories.

【Keywords】:

Machine Learning 48

46. An Online Spectral Learning Algorithm for Partially Observable Nonlinear Dynamical Systems.

Paper Link】 【Pages】:

【Authors】: Byron Boots ; Geoffrey J. Gordon

【Abstract】: Recently, a number of researchers have proposed spectral algorithms for learning models of dynamical systems — for example, Hidden Markov Models (HMMs), Partially Observable Markov Decision Processes (POMDPs), and Transformed Predictive State Representations (TPSRs). These algorithms are attractive since they are statistically consistent and not subject to local optima. However, they are batch methods: they need to store their entire training data set in memory at once and operate on it as a large matrix, and so they cannot scale to extremely large data sets (either many examples or many features per example). In turn, this restriction limits their ability to learn accurate models of complex systems. To overcome these limitations, we propose a new online spectral algorithm, which uses tricks such as incremental Singular Value Decomposition (SVD) and random projections to scale to much larger data sets and more complex systems than previous methods. We demonstrate the new method on an inertial measurement prediction task and a high-bandwidth video mapping task and we illustrate desirable behaviors such as "closing the loop," where the latent state representation changes suddenly as the learner recognizes that it has returned to a previously known place.

【Keywords】:

47. Learning Structured Embeddings of Knowledge Bases.

Paper Link】 【Pages】:

【Authors】: Antoine Bordes ; Jason Weston ; Ronan Collobert ; Yoshua Bengio

【Abstract】: Many Knowledge Bases (KBs) are now readily available and encompass colossal quantities of information thanks to either a long-term funding effort (e.g. WordNet, OpenCyc) or a collaborative process (e.g. Freebase, DBpedia). However, each of them is based on a different rigorous symbolic framework which makes it hard to use their data in other systems. It is unfortunate because such rich structured knowledge might lead to a huge leap forward in many other areas of AI like nat- ural language processing (word-sense disambiguation, natural language understanding, ...), vision (scene classification, image semantic annotation, ...) or collaborative filtering. In this paper, we present a learning process based on an innovative neural network architecture designed to embed any of these symbolic representations into a more flexible continuous vector space in which the original knowledge is kept and enhanced. These learnt embeddings would allow data from any KB to be easily used in recent machine learning meth- ods for prediction and information retrieval. We illustrate our method on WordNet and Freebase and also present a way to adapt it to knowledge extraction from raw text.

【Keywords】:

48. A Nonparametric Bayesian Model of Multi-Level Category Learning.

Paper Link】 【Pages】:

【Authors】: Kevin Robert Canini ; Thomas L. Griffiths

【Abstract】: Categories are often organized into hierarchical taxonomies, that is, tree structures where each node represents a labeled category, and a node's parent and children are, respectively, the category's supertype and subtypes. A natural question is whether it is possible to reconstruct category taxonomies in cases where we are not given explicit information about how categories are related to each other, but only a sample of observations of the members of each category. In this paper, we introduce a nonparametric Bayesian model of multi-level category learning, an extension of the hierarchical Dirichlet process (HDP) that we call the tree-HDP. We demonstrate the ability of the tree-HDP to reconstruct simulated datasets of artificial taxonomies, and show that it produces similar performance to human learners on a taxonomy inference task.

【Keywords】:

49. Large Scale Spectral Clustering with Landmark-Based Representation.

Paper Link】 【Pages】:

【Authors】: Xinlei Chen ; Deng Cai

【Abstract】: Spectral clustering is one of the most popular clustering approaches. Despite its good performance, it is limited in its applicability to large-scale problems due to its high computational complexity. Recently, many approaches have been proposed to accelerate the spectral clustering. Unfortunately, these methods usually sacrifice quite a lot information of the original data, thus result in a degradation of performance. In this paper, we propose a novel approach, called Landmark-based Spectral Clustering (LSC), for large scale clustering problems. Specifically, we select $p\ (\ll n)$ representative data points as the landmarks and represent the original data points as the linear combinations of these landmarks. The spectral embedding of the data can then be efficiently computed with the landmark-based representation. The proposed algorithm scales linearly with the problem size. Extensive experiments show the effectiveness and efficiency of our approach comparing to the state-of-the-art methods.

【Keywords】:

50. Unsupervised Learning of Human Behaviours.

Paper Link】 【Pages】:

【Authors】: Sook-Ling Chua ; Stephen Marsland ; Hans W. Guesgen

【Abstract】: Behaviour recognition is the process of inferring the behaviour of an individual from a series of observations acquired from sensors such as in a smart home. The majority of existing behaviour recognition systems are based on supervised learning algorithms, which means that training them requires a preprocessed, annotated dataset. Unfortunately, annotating a dataset is a rather tedious process and one that is prone to error. In this paper we suggest a way to identify structure in the data based on text compression and the edit distance between words, without any prior labelling. We demonstrate that by using this method we can automatically identify patterns and segment the data into patterns that correspond to human behaviours. To evaluate the effectiveness of our proposed method, we use a dataset from a smart home and compare the labels produced by our approach with the labels assigned by a human to the activities in the dataset. We find that the results are promising and show significant improvement in the recognition accuracy over Self-Organising Maps (SOMs).

【Keywords】:

51. Basis Function Discovery Using Spectral Clustering and Bisimulation Metrics.

Paper Link】 【Pages】:

【Authors】: Gheorghe Comanici ; Doina Precup

【Abstract】: We study the problem of automatically generating features for function approximation in reinforcement learning. We build on the work of Mahadevan and his colleagues, who pioneered the use of spectral clustering methods for basis function construction. Their methods work on top of a graph that captures state adjacency. Instead, we use bisimulation metrics in order to provide state distances for spectral clustering. The advantage of these metrics is that they incorporate reward information in a natural way, in addition to the state transition information. We provide theoretical bounds on the quality of the obtained approximation, which justify the importance of incorporating reward information. We also demonstrate empirically that the approximation quality improves when bisimulation metrics are used instead of the state adjacency graph in the basis function construction process.

【Keywords】:

52. Item-Level Social Influence Prediction with Probabilistic Hybrid Factor Matrix Factorization.

Paper Link】 【Pages】:

【Authors】: Peng Cui ; Fei Wang ; Shiqiang Yang ; Lifeng Sun

【Abstract】: Social influence has become the essential factor which drives the dynamic evolution process of social network structure and user behaviors. Previous research often focus on social influence analysis in network-level or topic-level. In this paper, we concentrate on predicting item-level social influence to reveal the users' influences in a more fine-grained level. We formulate the social influence prediction problem as the estimation of a user-post matrix, where each entry in the matrix represents the social influence strength the corresponding user has given the corresponding web post. To deal with the sparsity and complex factor challenges in the research, we model the problem by extending the probabilistic matrix factorization method to incorporate rich prior knowledge on both user dimension and web post dimension, and propose the Probabilistic Hybrid Factor Matrix Factorization (PHF-MF) approach. Intensive experiments are conducted on a real world online social network to demonstrate the advantages and characteristics of the proposed method.

【Keywords】:

53. Selective Transfer Between Learning Tasks Using Task-Based Boosting.

Paper Link】 【Pages】:

【Authors】: Eric Eaton ; Marie desJardins

【Abstract】: The success of transfer learning on a target task is highly dependent on the selected source data. Instance transfer methods reuse data from the source tasks to augment the training data for the target task. If poorly chosen, this source data may inhibit learning, resulting in negative transfer. The current most widely used algorithm for instance transfer, TrAdaBoost, performs poorly when given irrelevant source data. We present a novel task-based boosting technique for instance transfer that selectively chooses the source knowledge to transfer to the target task. Our approach performs boosting at both the instance level and the task level, assigning higher weight to those source tasks that show positive transferability to the target task, and adjusting the weights of individual instances within each source task via AdaBoost. We show that this combination of task- and instance-level boosting significantly improves transfer performance over existing instance transfer algorithms when given a mix of relevant and irrelevant source data, especially for small amounts of data on the target task.

【Keywords】:

54. Across-Model Collective Ensemble Classification.

Paper Link】 【Pages】:

【Authors】: Hoda Eldardiry ; Jennifer Neville

【Abstract】: Ensemble classification methods that independently construct component models (e.g., bagging) improve accuracy over single models by reducing the error due to variance. Some work has been done to extend ensemble techniques for classification in relational domains by taking relational data characteristics or multiple link types into account during model construction. However, since these approaches follow the conventional approach to ensemble learning, they improve performance by reducing the error due to variance in learning. We note however, that variance in inference can be an additional source of error in relational methods that use collective classification, since inferred values are propagated during inference. We propose a novel ensemble mechanism for collective classification that reduces  both learning and inference variance, by incorporating prediction averaging into the collective inference process itself. We show that our proposed method significantly outperforms a straightforward relational ensemble baseline on both synthetic and real-world datasets.

【Keywords】:

55. Symmetric Graph Regularized Constraint Propagation.

Paper Link】 【Pages】:

【Authors】: Zhenyong Fu ; Zhiwu Lu ; Horace Ho-Shing Ip ; Yuxin Peng ; Hongtao Lu

【Abstract】: This paper presents a novel symmetric graph regularization framework for pairwise constraint propagation. We first decompose the challenging problem of pairwise constraint propagation into a series of two-class label propagation subproblems and then deal with these subproblems by quadratic optimization with symmetric graph regularization. More importantly, we clearly show that pairwise constraint propagation is actually equivalent to solving a Lyapunov matrix equation, which is widely used in Control Theory as a standard continuous-time equation. Different from most previous constraint propagation methods that suffer from severe limitations, our method can directly be applied to multi-class problem and also can effectively exploit both must-link and cannot-link constraints. The propagated constraints are further used to adjust the similarity between data points so that they can be incorporated into subsequent clustering. The proposed method has been tested in clustering tasks on six real-life data sets and then shown to achieve significant improvements with respect to the state of the arts.

【Keywords】:

56. A Feasible Nonconvex Relaxation Approach to Feature Selection.

Paper Link】 【Pages】:

【Authors】: Cuixia Gao ; Naiyan Wang ; Qi Yu ; Zhihua Zhang

【Abstract】: Variable selection problems are typically addressed under apenalized optimization framework. Nonconvex penalties such as the minimax concave plus (MCP) and smoothly clipped absolute deviation(SCAD), have been demonstrated to have the properties of sparsity practically and theoretically. In this paper we propose a new nonconvex penalty that we call exponential-type penalty. The exponential-type penalty is characterized by a positive parameter,which establishes a connection with the ell 0 and ell 1 penalties.We apply this new penalty to sparse supervised learning problems. To solve to resulting optimization problem, we resort to a reweighted ell 1 minimization method. Moreover, we devise an efficient method for the adaptive update of the tuning parameter. Our experimental results are encouraging. They show that the exponential-type penalty is competitive with MCP and SCAD.

【Keywords】:

57. OASIS: Online Active Semi-Supervised Learning.

Paper Link】 【Pages】:

【Authors】: Andrew B. Goldberg ; Xiaojin Zhu ; Alex Furger ; Jun-Ming Xu

【Abstract】: We consider a learning setting of importance to large scale machine learning: potentially unlimited data arrives sequentially, but only a small fraction of it is labeled. The learner cannot store the data; it should learn from both labeled and unlabeled data, and it may also request labels for some of the unlabeled items. This setting is frequently encountered in real-world applications and has the characteristics of online, semi-supervised, and active learning. Yet previous learning models fail to consider these characteristics jointly. We present OASIS, a Bayesian model for this learning setting. The main contributions of the model include the novel integration of a semi-supervised likelihood function, a sequential Monte Carlo scheme for efficient online Bayesian updating, and a posterior-reduction criterion for active learning. Encouraging results on both synthetic and real-world optical character recognition data demonstrate the synergy of these characteristics in OASIS.

【Keywords】:

58. Learning a Kernel for Multi-Task Clustering.

Paper Link】 【Pages】:

【Authors】: Quanquan Gu ; Zhenhui Li ; Jiawei Han

【Abstract】: Multi-task learning has received increasing attention in the past decade. Many supervised multi-task learning methods have been proposed, while unsupervised multi-task learning is still a rarely studied problem. In this paper, we propose to learn a kernel for multi-task clustering. Our goal is to learn a Reproducing Kernel Hilbert Space, in which the geometric structure of the data in each task is preserved, while the data distributions of any two tasks are as close as possible. This is formulated as a unified kernel learning framework, under which we study two types of kernel learning: nonparametric kernel learning and spectral kernel design. Both types of kernel learning can be solved by linear programming. Experiments on several cross-domain text data sets demonstrate that kernel k-means on the learned kernel can achieve better clustering results than traditional single-task clustering methods. It also outperforms the newly proposed multi-task clustering method.

【Keywords】:

59. Adaptive Large Margin Training for Multilabel Classification.

Paper Link】 【Pages】:

【Authors】: Yuhong Guo ; Dale Schuurmans

【Abstract】: Multilabel classification is a central problem in many areas of data analysis, including text and multimedia categorization, where individual data objects need to be assigned multiple labels. A key challenge in these tasks is to learn a classifier that can properly exploit label correlations without requiring exponential enumeration of label subsets during training or testing. We investigate novel loss functions for multilabel training within a large margin framework---identifying a simple alternative that yields improved generalization while still allowing efficient training. We furthermore show how covariances between the label models can be learned simultaneously with the classification model itself, in a jointly convex formulation, without compromising scalability. The resulting combination yields state of the art accuracy in multilabel webpage classification.

【Keywords】:

60. Value Function Approximation in Reinforcement Learning Using the Fourier Basis.

Paper Link】 【Pages】:

【Authors】: George Konidaris ; Sarah Osentoski ; Philip S. Thomas

【Abstract】: We describe the Fourier basis, a linear value function approximation scheme based on the Fourier series. We empirically demonstrate that it performs well compared to radial basis functions and the polynomial basis, the two most popular fixed bases for linear value function approximation, and is competitive with learned proto-value functions.

【Keywords】:

61. Improving Semi-Supervised Support Vector Machines Through Unlabeled Instances Selection.

Paper Link】 【Pages】:

【Authors】: Yu-Feng Li ; Zhi-Hua Zhou

【Abstract】: Semi-supervised support vector machines (S3VMs) are a kind of popular approaches which try to improve learning performance by exploiting unlabeled data. Though S3VMs have been found helpful in many situations, they may degenerate performance and the resultant generalization ability may be even worse than using the labeled data only. In this paper, we try to reduce the chance of performance degeneration of S3VMs. Our basic idea is that, rather than exploiting all unlabeled data, the unlabeled instances should be selected such that only the ones which are very likely to be helpful are exploited, while some highly risky unlabeled instances are avoided. We propose the S3VM- us method by using hierarchical clustering to select the unlabeled instances. Experiments on a broad range of data sets over eighty-eight different settings show that the chance of performance degeneration of S3VM- us is much smaller than that of existing S3VMs.

【Keywords】:

62. Size Adaptive Selection of Most Informative Features.

Paper Link】 【Pages】:

【Authors】: Si Liu ; Hairong Liu ; Longin Jan Latecki ; Shuicheng Yan ; Changsheng Xu ; Hanqing Lu

【Abstract】: In this paper, we propose a novel method to select the most informativesubset of features, which has little redundancy andvery strong discriminating power. Our proposed approach automaticallydetermines the optimal number of features and selectsthe best subset accordingly by maximizing the averagepairwise informativeness, thus has obvious advantage overtraditional filter methods. By relaxing the essential combinatorialoptimization problem into the standard quadratic programmingproblem, the most informative feature subset canbe obtained efficiently, and a strategy to dynamically computethe redundancy between feature pairs further greatly acceleratesour method through avoiding unnecessary computationsof mutual information. As shown by the extensive experiments,the proposed method can successfully select the mostinformative subset of features, and the obtained classificationresults significantly outperform the state-of-the-art results onmost test datasets.

【Keywords】:

63. Ordinal Regression via Manifold Learning.

Paper Link】 【Pages】:

【Authors】: Yang Liu ; Yan Liu ; Keith C. C. Chan

【Abstract】: Ordinal regression is an important research topic in machine learning. It aims to automatically determine the implied rating of a data item on a fixed, discrete rating scale. In this paper, we present a novel ordinal regression approach via manifold learning, which is capable of uncovering the embedded nonlinear structure of the data set according to the observations in the highdimensional feature space. By optimizing the order information of the observations and preserving the intrinsic geometry of the data set simultaneously, the proposed algorithm provides the faithful ordinal regression to the new coming data points. To offer more general solution to the data with natural tensor structure, we further introduce the multilinear extension of the proposed algorithm, which can support the ordinal regression of high order data like images. Experiments on various data sets validate the effectiveness of the proposed algorithm as well as its extension.

【Keywords】:

64. Mean Field Inference in Dependency Networks: An Empirical Study.

Paper Link】 【Pages】:

【Authors】: Daniel Lowd ; Arash Shamaei

【Abstract】: Dependency networks are a compelling alternative to Bayesian networks for learning joint probability distributions from data and using them to compute probabilities. A dependency network consists of a set of conditional probability distributions, each representing the probability of a single variable given its Markov blanket. Running Gibbs sampling with these conditional distributions produces a joint distribution that can be used to answer queries, but suffers from the traditional slowness of sampling-based inference. In this paper, we observe that the mean field update equation can be applied to dependency networks, even though the conditional probability distributions may be inconsistent with each other. In experiments with learning and inference on 12 datasets, we demonstrate that mean field inference in dependency networks offers similar accuracy to Gibbs sampling but with orders of magnitude improvements in speed. Compared to Bayesian networks learned on the same data, dependency networks offer higher accuracy at greater amounts of evidence. Furthermore, mean field inference is consistently more accurate in dependency networks than in Bayesian networks learned on the same data.

【Keywords】:

65. Latent Semantic Learning by Efficient Sparse Coding with Hypergraph Regularization.

Paper Link】 【Pages】:

【Authors】: Zhiwu Lu ; Yuxin Peng

【Abstract】: This paper presents a novel latent semantic learning algorithm for action recognition. Through efficient sparse coding, we can learn latent semantics (i.e. high-level features) from a large vocabulary of abundant mid-level features (i.e. visual keywords). More importantly, we can capture the manifold structure hidden among mid-level features by incorporating hypergraph regularization into sparse coding. The learnt latent semantics can further be readily used for action recognition by defining a histogram intersection kernel. Different from the traditional latent semantic analysis based on topic models, our sparse coding method with hypergraph regularization can exploit the manifold structure hidden among mid-level features for latent semantic learning, which results in compact but discriminative high-level features for action recognition. We have tested our method on the commonly used KTH action dataset and the unconstrained YouTube action dataset. The experimental results show the superior performance of our method.

【Keywords】:

66. Linear Discriminant Analysis: New Formulations and Overfit Analysis.

Paper Link】 【Pages】:

【Authors】: Dijun Luo ; Chris H. Q. Ding ; Heng Huang

【Abstract】: In this paper, we will present a unified view for LDA. We will (1) emphasize that standard LDA solutions are not unique, (2) propose several new LDA formulations: St-orthonormal LDA, Sw-orthonormal LDA and orthogonal LDA which have unique solutions, and (3) show that with St-orthonormal LDA and Sw-orthonormal LDA formulations, solutions to all four major LDA objective functions are identical. Furthermore, we perform an indepth analysis to show that the LDA sometimes performs poorly due to over-fitting, i.e., it picks up PCA dimensions with small eigenvalues. From this analysis, we propose a stable LDA which uses PCA first to reduce to a small PCA subspace and do LDA in the subspace.

【Keywords】:

67. Multi-Level Cluster Indicator Decompositions of Matrices and Tensors.

Paper Link】 【Pages】:

【Authors】: Dijun Luo ; Chris H. Q. Ding ; Heng Huang

【Abstract】: A main challenging problem for many machine learning and data mining applications is that the amount of data and features are very large, so that low-rank approximations of original data are often required for efficient computation. We propose new multi-level clustering based low-rank matrix approximations which are comparable and even more compact than Singular Value Decomposition (SVD). We utilize the cluster indicators of data clustering results to form the subspaces, hence our decomposition results are more interpretable. We further generalize our clustering based matrix decompositions to tensor decompositions that are useful in high-order data analysis. We also provide an upper bound for the approximation error of our tensor decomposition algorithm. In all experimental results, our methods significantly outperform traditional decomposition methods such as SVD and high-order SVD.

【Keywords】:

68. Sparse Group Restricted Boltzmann Machines.

Paper Link】 【Pages】:

【Authors】: Heng Luo ; Ruimin Shen ; Changyong Niu ; Carsten Ullrich

【Abstract】: Since learning in Boltzmann machines is typically quite slow, there is a need to restrict connections within hidden layers. However, theresulting states of hidden units exhibit statistical dependencies. Based on this observation, we propose using l1/l2 regularization upon the activation probabilities of hidden units in restricted Boltzmann machines to capture the local dependencies among hidden units. This regularization not only encourages hidden units of many groups to be inactive given observed data but also makes hidden units within a group compete with each other for modeling observed data. Thus, the l1/l2 regularization on RBMs yields sparsity at both the group and the hidden unit levels. We call RBMs trained with the regularizer sparse group RBMs (SGRBMs). The proposed SGRBMs are appliedto model patches of natural images, handwritten digits and OCR English letters. Then to emphasize that SGRBMs can learn more discriminative features we applied SGRBMs to pretrain deep networks for classification tasks. Furthermore, we illustrate the regularizer can also be applied to deep Boltzmann machines, which lead to sparse group deep Boltzmann machines. When adapted to the MNIST data set, a two-layer sparse group Boltzmann machine achieves an error rate of 0.84%, which is, to our knowledge, the best published result on the permutation-invariant version of the MNIST task.

【Keywords】:

69. Scaling Up Reinforcement Learning through Targeted Exploration.

Paper Link】 【Pages】:

【Authors】: Timothy Arthur Mann ; Yoonsuck Choe

【Abstract】: Recent Reinforcement Learning (RL) algorithms, such as R-MAX, make (with high probability) only a small number of poor decisions. In practice, these algorithms do not scale well as the number of states grows because the algorithms spend too much effort exploring. We introduce an RL algorithm State TArgeted R-MAX (STAR-MAX) that explores a subset of the state space, called the exploration envelope ξ. When ξ equals the total state space, STAR-MAX behaves identically to R-MAX. When ξ is a subset of the state space, to keep exploration within ξ, a recovery rule β is needed. We compared existing algorithms with our algorithm employing various exploration envelopes. With an appropriate choice of ξ, STAR-MAX scales far better than existing RL algorithms as the number of states increases. A possible drawback of our algorithm is its dependence on a good choice of ξ and β. However, we show that an effective recovery rule β can be learned on-line and ξ can be learned from demonstrations. We also find that even randomly sampled exploration envelopes can improve cumulative rewards compared to R-MAX. We expect these results to lead to more efficient methods for RL in large-scale problems.

【Keywords】:

70. Differential Eligibility Vectors for Advantage Updating and Gradient Methods.

Paper Link】 【Pages】:

【Authors】: Francisco S. Melo

【Abstract】: In this paper we propose differential eligibility vectors (DEV) for temporal-difference (TD) learning, a new class of eligibility vectors designed to bring out the contribution of each action in the TD-error at each state. Specifically, we use DEV in TD-Q(lambda) to more accurately learn the relative value of the actions, rather than their absolute value. We identify conditions that ensure convergence w.p.1 of TD-Q(lambda) with DEV and show that this algorithm can also be used to directly approximate the advantage function associated with a given policy, without the need to compute an auxiliary function - something that, to the extent of our knowledge, was not known possible. Finally, we discuss the integration of DEV in LSTDQ and actor-critic algorithms.

【Keywords】:

71. Markov Logic Sets: Towards Lifted Information Retrieval Using PageRank and Label Propagation.

Paper Link】 【Pages】:

【Authors】: Marion Neumann ; Babak Ahmadi ; Kristian Kersting

【Abstract】: Inspired by “GoogleTM Sets” and Bayesian sets, we consider the problem of retrieving complex objects and relations among them, i.e., ground atoms from a logical concept, given a query consisting of a few atoms from that concept. We formulate this as a within-network relational learning problem using few labels only and describe an algorithm that ranks atoms using a score based on random walks with restart (RWR): the probability that a random surfer hits an atom starting from the query atoms. Specifically, we compute an initial ranking using personalized PageRank. Then, we find paths of atoms that are connected via their arguments, variablize the ground atoms in each path, in order to create features for the query. These features are used to re-personalize the original RWR and to finally compute the set completion, based on Label Propagation. Moreover, we exploit that RWR techniques can naturally be lifted and show that lifted inference for label propagation is possible. We evaluate our algorithm on a realworld relational dataset by finding completions of sets of objects describing the Roman city of Pompeii. We compare to Bayesian sets and show that our approach gives very reasonable set completions.

【Keywords】:

72. Efficiently Learning a Distance Metric for Large Margin Nearest Neighbor Classification.

Paper Link】 【Pages】:

【Authors】: Kyoungup Park ; Chunhua Shen ; Zhihui Hao ; Junae Kim

【Abstract】: We concern the problem of learning a Mahalanobis distance metric for improving nearest neighbor classification. Our work is built upon the large margin nearest neighbor (LMNN) classification framework. Due to the semidefiniteness constraint in the optimization problem of LMNN, it is not scalable in terms of the dimensionality of the input data. The original LMNN solver partially alleviates this problem by adopting alternating projection methods instead of standard interior-point methods. Still, at each iteration, the computation complexity is at least O(D 3 ) (D is the dimension of input data). In this work, we propose a column generation based algorithm to solve the LMNN optimization problem much more efficiently. Our algorithm is much more scalable in tha tat each iteration, it does not need full eigen-decomposition. Instead, we only need to find the leading eigen value and its corresponding eigen vector, which is of O(D 2 ) complexity. Experiments show the efficiency and efficacy of our algorithms.

【Keywords】:

73. Non-Parametric Approximate Linear Programming for MDPs.

Paper Link】 【Pages】:

【Authors】: Jason Pazis ; Ronald Parr

【Abstract】: The Approximate Linear Programming (ALP) approach to value function approximation for MDPs is a parametric value function approximation method, in that it represents the value function as a linear combination of features which are chosen a priori. Choosing these features can be a difficult challenge in itself. One recent effort, Regularized Approximate Linear Programming (RALP), uses L1 regularization to address this issue by combining a large initial set of features with a regularization penalty that favors a smooth value function with few non-zero weights. Rather than using smoothness as a backhanded way of addressing the feature selection problem, this paper starts with smoothness and develops a non-parametric approach to ALP that is consistent with the smoothness assumption. We show that this new approach has some favorable practical and analytical properties in comparison to (R)ALP.

【Keywords】:

74. Optimal Rewards versus Leaf-Evaluation Heuristics in Planning Agents.

Paper Link】 【Pages】:

【Authors】: Jonathan Sorg ; Satinder P. Singh ; Richard L. Lewis

【Abstract】: Planning agents often lack the computational resources needed to build full planning trees for their environments. Agent designers commonly overcome this finite-horizon approximation by applying an evaluation function at the leaf-states of the planning tree. Recent work has proposed an alternative approach for overcoming computational constraints on agent design: modify the reward function. In this work, we compare this reward design approach to the common leaf-evaluation heuristic approach for improving planning agents. We show that in many agents, the reward design approach strictly subsumes the leaf-evaluation approach, i.e., there exists a reward function for every leaf-evaluation heuristic that leads to equivalent behavior, but the converse is not true. We demonstrate that this generality leads to improved performance when an agent makes approximations in addition to the finite-horizon approximation. As part of our contribution, we extend PGRD, an online reward design algorithm, to develop reward design algorithms for Sparse Sampling and UCT, two algorithms capable of planning in large state spaces.

【Keywords】:

75. A Generalised Solution to the Out-of-Sample Extension Problem in Manifold Learning.

Paper Link】 【Pages】:

【Authors】: Harry Strange ; Reyer Zwiggelaar

【Abstract】: Manifold learning is a powerful tool for reducing the dimensionality of a dataset by finding a low-dimensional embedding that retains important geometric and topological features. In many applications it is desirable to add new samples to a previously learnt embedding, this process of adding new samples is known as the out-of-sample extension problem. Since many manifold learning algorithms do not naturally allow for new samples to be added we present an easy to implement generalized solution to the problem that can be used with any existing manifold learning algorithm. Our algorithm is based on simple geometric intuition about the local structure of a manifold and our results show that it can be effectively used to add new samples to a previously learnt embedding. We test our algorithm on both artificial and real world image data and show that our method significantly out performs existing out-of-sample extension strategies.

【Keywords】:

76. Collaborative Users' Brand Preference Mining across Multiple Domains from Implicit Feedbacks.

Paper Link】 【Pages】:

【Authors】: Jian Tang ; Jun Yan ; Lei Ji ; Ming Zhang ; Shaodan Guo ; Ning Liu ; Xianfang Wang ; Zheng Chen

【Abstract】: Advanced e-applications require comprehensive knowledge about their users’ preferences in order to provide accurate personalized services. In this paper, we propose to learn users’ preferences to product brands from their implicit feedbacks such as their searching and browsing behaviors in user Web browsing log data. The user brand preference learning problem is challenge since (1) the users’ implicit feedbacks are extremely sparse in various product domains; and (2) we can only observe positive feedbacks from users’ behaviors. In this paper, we propose a latent factor model to collaboratively mine users’ brand preferences across multiple domains simultaneously. By collective learning, the learning processes in all the domains are mutually enhanced and hence the problem of data scarcity in each single domain can be effectively addressed. On the other hand, we learn our model with an adaption of the Bayesian personalized ranking (BPR) optimization criterion which is a general learning framework for collaborative filtering from implicit feedbacks. Experiments with both synthetic and real world datasets show that our proposed model significantly outperforms the baselines.

【Keywords】:

77. Towards Maximizing the Area Under the ROC Curve for Multi-Class Classification Problems.

Paper Link】 【Pages】:

【Authors】: Ke Tang ; Rui Wang ; Tianshi Chen

【Abstract】: The Area Under the ROC Curve (AUC) metric has achieved a big success in binary classification problems since they measure the performance of classifiers without making any specific assumptions about the class distribution and misclassification costs. This is desirable because the class distribution and misclassification costs may be unknown during training process or even change in environment. MAUC, the extension of AUC to multi-class problems, has also attracted a lot of attention. However, despite the emergence of approaches for training classifiers with large AUC, little has been done for MAUC. This paper analyzes MAUC in-depth, and reveals that the maximization of MAUC can be achieved by decomposing the multi-class problem into a number of independent sub-problems. These sub-problems are formulated in the form of a “learning to rank” problem, for which well-established methods already exist. Based on the analysis, a method that employs RankBoost algorithm as the sub-problem solver is proposed to achieve classification systems with maximum MAUC. Empirical studies have shown the advantages of the proposed method over other eight relevant methods. Due to the importance of MAUC to multi-class cost-sensitive learning and class imbalanced learning problems, the proposed method is a general technique for both problems. It can also be generalized to accommodate other learning algorithms as the sub-problem solvers.

【Keywords】:

78. Fast Newton-CG Method for Batch Learning of Conditional Random Fields.

Paper Link】 【Pages】:

【Authors】: Yuta Tsuboi ; Yuya Unno ; Hisashi Kashima ; Naoaki Okazaki

【Abstract】: We propose a fast batch learning method for linear-chain Conditional Random Fields (CRFs) based on Newton-CG methods. Newton-CG methods are a variant of Newton method for high-dimensional problems. They only require the Hessian-vector products instead of the full Hessian matrices. To speed up Newton-CG methods for the CRF learning, we derive a novel dynamic programming procedure for the Hessian-vector products of the CRF objective function. The proposed procedure can reuse the byproducts of the time-consuming gradient computation for the Hessian-vector products to drastically reduce the total computation time of the Newton-CG methods. In experiments with tasks in natural language processing, the proposed method outperforms a conventional quasi-Newton method. Remarkably, the proposed method is competitive with online learning algorithms that are fast but unstable.

【Keywords】:

79. Automatic Group Sparse Coding.

Paper Link】 【Pages】:

【Authors】: Fei Wang ; Noah Lee ; Jimeng Sun ; Jianying Hu ; Shahram Ebadollahi

【Abstract】: Sparse Coding (SC), which models the data vectors as sparse linear combinations over basis vectors (i.e., dictionary), has been widely applied in machine learning, signal processing and neuroscience. Recently, one specific SC technique, Group Sparse Coding (GSC), has been proposed to learn a common dictionary over multiple different groups of data, where the data groups are assumed to be pre-defined. In practice, this may not always be the case. In this paper, we propose Automatic Group Sparse Coding (AutoGSC), which can (1) discover the hidden data groups; (2) learn a common dictionary over different data groups; and (3) learn an individual dictionary for each data group. Finally, we conduct experiments on both synthetic and real world data sets to demonstrate the effectiveness of AutoGSC, and compare it with traditional sparse coding and Nonnegative Matrix Factorization (NMF) methods.

【Keywords】:

80. Towards Evolutionary Nonnegative Matrix Factorization.

Paper Link】 【Pages】:

【Authors】: Fei Wang ; Hanghang Tong ; Ching-Yung Lin

【Abstract】: Nonnegative Matrix Factorization (NMF) techniques has aroused considerable interests from the field of artificial intelligence in recent years because of its good interpretability and computational efficiency. However, in many real world applications, the data features usually evolve over time smoothly. In this case, it would be very expensive in both computation and storage to rerun the whole NMF procedure after each time when the data feature changing. In this paper, we propose Evolutionary Nonnegative Matrix Factorization (eNMF), which aims to incrementally update the factorized matrices in a computation and space efficient manner with the variation of the data matrix. We devise such evolutionary procedure for both asymmetric and symmetric NMF. Finally we conduct experiments on several real world data sets to demonstrate the efficacy and efficiency of eNMF.

【Keywords】:

81. Learning Instance Specific Distance for Multi-Instance Classification.

Paper Link】 【Pages】:

【Authors】: Hua Wang ; Feiping Nie ; Heng Huang

【Abstract】: Multi-Instance Learning (MIL) deals with problems where each training example is a bag, and each bag contains a set of instances. Multi-instance representation is useful in many real world applications, because it is able to capture more structural information than traditional flat single-instance representation. However, it also brings new challenges. Specifically, the distance between data objects in MIL is a set-to-set distance, which is harder to estimate than vector distances used in single-instance data. Moreover, because in MIL labels are assigned to bags instead of instances, although a bag belongs to a class, some, or even most, of its instances may not be truly related to the class. In order to address these difficulties, in this paper we propose a novel Instance Specific Distance (ISD) method for MIL, which computes the Class-to-Bag (C2B) distance by further considering the relevances of training instances with respect to their labeled classes. Taking into account the outliers caused by the weak label association in MIL, we learn ISD by solving an l0+-norm minimization problem. An efficient algorithm to solve the optimization problem is presented, together with the rigorous proof of its convergence. The promising results on five benchmark multi-instance data sets and two real world multi-instance applications validate the effectiveness of the proposed method.

【Keywords】:

82. Transfer Learning by Structural Analogy.

Paper Link】 【Pages】:

【Authors】: Hua-Yan Wang ; Qiang Yang

【Abstract】: Transfer learning allows knowledge to be extracted from auxiliary domains and be used to enhance learning in a target domain. For transfer learning to be successful, it is critical to find the similarity between auxiliary and target domains, even when such mappings are not obvious. In this paper, we present a novel algorithm for finding the structural similarity between two domains, to enable transfer learning at a structured knowledge level. In particular, we address the problem of how to learn a non-trivial structural similarity mapping between two different domains when they are completely different on the representation level. This problem is challenging because we cannot directly compare features across domains. Our algorithm extracts the structural features within each domain and then maps the features into the Reproducing Kernel Hilbert Space (RKHS), such that the "structural dependencies" of features across domains can be estimated by kernel matrices of the features within each domain. By treating the analogues from both domains as equivalent, we can transfer knowledge to achieve a better understanding of the domains and improved performance for learning. We validate our approach on synthetic and real-world datasets.

【Keywords】:

83. Efficient Subspace Segmentation via Quadratic Programming.

Paper Link】 【Pages】:

【Authors】: Shusen Wang ; Xiaotong Yuan ; Tiansheng Yao ; Shuicheng Yan ; Jialie Shen

【Abstract】: We explore in this paper efficient algorithmic solutions to robustsubspace segmentation. We propose the SSQP, namely SubspaceSegmentation via Quadratic Programming, to partition data drawnfrom multiple subspaces into multiple clusters. The basic idea ofSSQP is to express each datum as the linear combination of otherdata regularized by an overall term targeting zero reconstructioncoefficients over vectors from different subspaces. The derivedcoefficient matrix by solving a quadratic programming problem istaken as an affinity matrix, upon which spectral clustering isapplied to obtain the ultimate segmentation result. Similar tosparse subspace clustering (SCC) and low-rank representation (LRR),SSQP is robust to data noises as validated by experiments on toydata. Experiments on Hopkins 155 database show that SSQP can achievecompetitive accuracy as SCC and LRR in segmenting affine subspaces,while experimental results on the Extended Yale Face Database Bdemonstrate SSQP's superiority over SCC and LRR. Beyond segmentationaccuracy, all experiments show that SSQP is much faster than bothSSC and LRR in the practice of subspace segmentation.

【Keywords】:

84. Localized K-Flats.

Paper Link】 【Pages】:

【Authors】: Yong Wang ; Yuan Jiang ; Yi Wu ; Zhi-Hua Zhou

【Abstract】: K-flats is a model-based linear manifold clustering algorithm which has been successfully applied in many real-world scenarios. Though some previous works have shown that K-flats doesn’t always provide good performance, little effort has been devoted to analyze its inherent deficiency. In this paper, we address this challenge by showing that the deteriorative performance of K-flats can be attributed to the usual reconstruction error measure and the infinitely extending representations of linear models. Then we propose Localized K-flats algorithm (LKF), which introduces localized representations of linear models and a new distortion measure, to remove confusion among different clusters. Experiments on both synthetic and real-world data sets demonstrate the efficiency of the proposed algorithm. Moreover, preliminary experiments show that LKF has the potential to group manifolds with nonlinear structure.

【Keywords】:

85. Heterogeneous Transfer Learning with RBMs.

Paper Link】 【Pages】:

【Authors】: Bin Wei ; Christopher J. Pal

【Abstract】: A common approach in machine learning is to use a large amount of labeled data to train a model. Usually this model can then only be used to classify data in the same feature space. However, labeled data is often expensive to obtain. A number of strategies have been developed by the machine learning community in recent years to address this problem, including: semi-supervised learning,domain adaptation,multi-task learning,and self-taught learning. While training data and test may have different distributions, they must remain in the same feature set. Furthermore, all the above methods work in the same feature space. In this paper, we consider an extreme case of transfer learning called heterogeneous transfer learning — where the feature spaces of the source task and the target tasks are disjoint. Previous approaches mostly fall in the multi-view learning category, where co-occurrence data from both feature spaces is required. We generalize the previous work on cross-lingual adaptation and propose a multi-task strategy for the task. We also propose the use of a restricted Boltzmann machine (RBM), a special type of probabilistic graphical models, as an implementation. We present experiments on two tasks: action recognition and cross-lingual sentiment classification.

【Keywords】:

86. Multi-Task Learning in Square Integrable Space.

Paper Link】 【Pages】:

【Authors】: Wei Wu ; Hang Li ; Yunhua Hu ; Rong Jin

【Abstract】: Several kernel based methods for multi-task learning have been proposed, which leverage relations among tasks as regularization to enhance the overall learning accuracies. These methods assume that the tasks share the same kernel, which could limit their applications because in practice different tasks may need different kernels. The main challenge of introducing multiple kernels into multiple tasks is that models from different Reproducing Kernel Hilbert Spaces (RKHSs) are not comparable, making it difficult to exploit relations among tasks. This paper addresses the challenge by formalizing the problem in the Square Integrable Space (SIS). Specially, it proposes a kernel based method which makes use of a regularization term defined in the SIS to represent task relations. We prove a new representer theorem for the proposed approach in SIS. We further derive a practical method for solving the learning problem and conduct consistency analysis of the method. We discuss the relations between our method and an existing method. We also give an SVM based implementation of our method for multi-label classification. Experiments on two real-world data sets show that the proposed method performs better than the existing method.

【Keywords】:

87. Sparse Matrix-Variate t Process Blockmodels.

Paper Link】 【Pages】:

【Authors】: Zenglin Xu ; Feng Yan ; Yuan Qi

【Abstract】: We consider the problem of modeling network interactions and identifying latent groups of network nodes. This problem is challenging due to the facts i) that the network nodes are interdependent instead of independent, ii) that the network data are very noisy (e.g., missing edges), and iii) that the network interactions are often sparse. To address these challenges, we propose a Sparse Matrix-variate t process Blockmodel (SMTB). In particular, we generalize a matrix-variate t distribution to a t process on matrices with nonlinear covariance functions. Due to this generalization, our model can estimate latent memberships for individual network nodes. This separates our model from previous t distribution based relational models. Also, we introduce sparse prior distributions on the latent membership parameters to select group assignments for individual nodes. To learn the model efficiently from data, we develop a variational method. When compared with several state-of-the-art models, including the predictive matrix-variate t models and mixed membership stochastic blockmodels, our model achieved improved prediction accuracy on real world network datasets.

【Keywords】:

88. Direct Density-Ratio Estimation with Dimensionality Reduction via Hetero-Distributional Subspace Analysis.

Paper Link】 【Pages】:

【Authors】: Makoto Yamada ; Masashi Sugiyama

【Abstract】: Methods for estimating the ratio of two probability density functions have been actively explored recently since they can be used for various data processing tasks such as non-stationarity adaptation, outlier detection, feature selection, and conditional probability estimation. In this paper, we propose a new density-ratio estimator which incorporates dimensionality reduction into the density-ratio estimation procedure. Through experiments, the proposed method is shown to compare favorably with existing density-ratio estimators in terms of both accuracy and computational costs.

【Keywords】:

89. Nonnegative Spectral Clustering with Discriminative Regularization.

Paper Link】 【Pages】:

【Authors】: Yi Yang ; Heng Tao Shen ; Feiping Nie ; Rongrong Ji ; Xiaofang Zhou

【Abstract】: Clustering is a fundamental research topic in the field of data mining. Optimizing the objective functions of clustering algorithms, e.g. normalized cut and k-means, is an NP-hard optimization problem. Existing algorithms usually relax the elements of cluster indicator matrix from discrete values to continuous ones. Eigenvalue decomposition is then performed to obtain a relaxed continuous solution, which must be discretized. The main problem is that the signs of the relaxed continuous solution are mixed. Such results may deviate severely from the true solution, making it a nontrivial task to get the cluster labels. To address the problem, we impose an explicit nonnegative constraint for a more accurate solution during the relaxation. Besides, we additionally introduce a discriminative regularization into the objective to avoid overfitting. A new iterative approach is proposed to optimize the objective. We show that the algorithm is a general one which naturally leads to other extensions. Experiments demonstrate the effectiveness of our algorithm.

【Keywords】:

90. Transfer Latent Semantic Learning: Microblog Mining with Less Supervision.

Paper Link】 【Pages】:

【Authors】: Dan Zhang ; Yan Liu ; Richard D. Lawrence ; Vijil Chenthamarakshan

【Abstract】: The increasing volume of information generated on micro-blogging sites such as Twitter raises several challenges to traditional text mining techniques. First, most texts from those sites are abbreviated due to the constraints of limited characters in one post; second, the input usually comes in streams of large-volumes. Therefore, it is of significant importance to develop effective and efficient representations of abbreviated texts for better filtering and mining. In this paper, we introduce a novel transfer learning approach, namely transfer latent semantic learning, that utilizes a large number of related tagged documents with rich information from other sources (source domain) to help build a robust latent semantic space for the abbreviated texts (target domain). This is achieved by simultaneously minimizing the document reconstruction error and the classification error of the labeled examples from the source domain by building a classifier with hinge loss in the latent semantic space. We demonstrate the effectiveness of our method by applying them to the task of classifying and tagging abbreviated texts. Experimental results on both synthetic datasets and real application datasets, including Reuters-21578 and Twitter data, suggest substantial improvements using our approach over existing ones.

【Keywords】:

91. Convex Sparse Coding, Subspace Learning, and Semi-Supervised Extensions.

Paper Link】 【Pages】:

【Authors】: Xinhua Zhang ; Yaoliang Yu ; Martha White ; Ruitong Huang ; Dale Schuurmans

【Abstract】: Automated feature discovery is a fundamental problem in machine learning. Although classical feature discovery methods do not guarantee optimal solutions in general, it has been recently noted that certain subspace learning and sparse coding problems can be solved efficiently, provided the number of features is not restricted a priori. We provide an extended characterization of this optimality result and describe the nature of the solutions under an expanded set of practical contexts. In particular, we apply the framework to a semi-supervised learning problem, and demonstrate that feature discovery can co-occur with input reconstruction and supervised training while still admitting globally optimal solutions. A comparison to existing semi-supervised feature discovery methods shows improved generalization and efficiency.

【Keywords】:

92. Multi-Task Learning in Heterogeneous Feature Spaces.

Paper Link】 【Pages】:

【Authors】: Yu Zhang ; Dit-Yan Yeung

【Abstract】: Multi-task learning aims at improving the generalization performance of a learning task with the help of some other related tasks. Although many multi-task learning methods have been proposed, they are all based on the assumption that all tasks share the same data representation. This assumption is too restrictive for general applications. In this paper, we propose a multi-task extension of linear discriminant analysis (LDA), called multi-task discriminant analysis (MTDA), which can deal with learning tasks with different data representations. For each task, MTDA learns a separate transformation which consists of two parts, one specific to the task and one common to all tasks. A by-product of MTDA is that it can alleviate the labeled data deficiency problem of LDA. Moreover, unlike many existing multi-task learning methods, MTDA can handle binary and multi-class problems for each task in a generic way. Experimental results on face recognition show that MTDA consistently outperforms related methods.

【Keywords】:

93. A Fast Spectral Relaxation Approach to Matrix Completion via Kronecker Products.

Paper Link】 【Pages】:

【Authors】: Hui Zhao ; Jiuqiang Han ; Naiyan Wang ; Congfu Xu ; Zhihua Zhang

【Abstract】: In the existing methods for solving matrix completion, such as singular value thresholding (SVT), soft-impute and fixed point continuation (FPCA) algorithms, it is typically required to repeatedly implement singular value decompositions (SVD) of matrices.When the size of the matrix in question is large, the computational complexity of finding a solution is costly. To reduce this expensive computational complexity, we apply Kronecker products to handle the matrix completion problem. In particular, we propose using Kronecker factorization, which approximates a matrix by the Kronecker product of several matrices of smaller sizes. Weintroduce Kronecker factorization into the soft-impute framework and devise an effective matrix completion algorithm.Especially when the factorized matrices have about the samesizes, the computational complexity of our algorithm is improved substantially.

【Keywords】:

Multiagent Systems 29

94. Refinement of Strong Stackelberg Equilibria in Security Games.

Paper Link】 【Pages】:

【Authors】: Bo An ; Milind Tambe ; Fernando Ordóñez ; Eric Anyung Shieh ; Christopher Kiekintveld

【Abstract】: Given the real-world deployments of attacker-defender Stackelberg security games, robustness to deviations from expected attacker behaviors has now emerged as a critically important issue. This paper provides four key contributions in this context. First, it identifies a fundamentally problematic aspect of current algorithms for security games. It shows that there are many situations where these algorithms face multiple equilibria, and they arbitrarily select one that may hand the defender a significant disadvantage, particularly if the attacker deviates from its equilibrium strategies due to unknown constraints. Second, for important subclasses of security games, it identifies situations where we will face such multiple equilibria. Third, to address these problematic situations, it presents two equilibrium refinement algorithms that can optimize the defender's utility if the attacker deviates from equilibrium strategies. Finally, it experimentally illustrates that the refinement approach achieved significant robustness in consideration of attackers' deviation due to unknown constraints.

【Keywords】:

95. Strategic Information Disclosure to People with Multiple Alternatives.

Paper Link】 【Pages】:

【Authors】: Amos Azaria ; Zinovi Rabinovich ; Sarit Kraus ; Claudia V. Goldman

【Abstract】: This paper studies how automated agents can persuade humans to behave in certain ways. The motivation behind such agent's behavior resides in the utility function that the agent's designer wants to maximize and which may be different from the user's utility function. Specifically, in the strategic settings studied, the agent provides correct yet partial information about a state of the world that is unknown to the user but relevant to his decision. Persuasion games were designed to study interactions between automated players where one player sends state information to the other to persuade it to behave in a certain way. We show that this game theory based model is not sufficient to model human-agent interactions, since people tend to deviate from the rational choice. We use machine learning to model such deviation in people from this game theory based model. The agent generates a probabilistic description of the world state that maximizes its benefit and presents it to the users. The proposed model was evaluated in an extensive empirical study involving road selection tasks that differ in length, costs and congestion. Results showed that people's behavior indeed deviated significantly from the behavior predicted by the game theory based model. Moreover, the agent developed in our model performed better than an agent that followed the behavior dictated by the game-theoretical models.

【Keywords】:

96. Branch and Price for Multi-Agent Plan Recognition.

Paper Link】 【Pages】:

【Authors】: Bikramjit Banerjee ; Landon Kraemer

【Abstract】: The problem of identifying the (dynamic) team structures and team behaviors from the observed activities of multiple agents is called Multi-Agent Plan Recognition (MAPR). We extend a recent formalization of this problem to accommodate a compact, partially ordered, multi-agent plan language, as well as complex plan execution models — particularly plan abandonment and activity interleaving. We adopt a branch and price approach to solve MAPR in such a challenging setting, and fully instantiate the (generic) pricing problem for MAPR. We show experimentally that this approach outperforms a recently proposed hypothesis pruning algorithm in two domains: multi-agent blocks word, and intrusion detection. The key benefit of the branch and price approach is its ability to grow the necessary component (occurrence) space from which the hypotheses are constructed, rather than begin with a fully enumerated component space that has an intractable size, and search it with pruning. Our formulation of MAPR has the broad objective of bringing mature Operations Research methodologies to bear upon MAPR, envisaged to have a similar impact as mature SAT-solvers had on planning.

【Keywords】:

Paper Link】 【Pages】:

【Authors】: Sofia Ceppi ; Nicola Gatti ; Enrico H. Gerding

【Abstract】: Recently there is an increase in smaller, domain-specific search engines that scour the deep web finding information that general-purpose engines are unable to discover. These search engines play a crucial role in the new generation of search paradigms where federated search engines (FSEs) integrate search results from heterogeneous sources. In this paper we pose, for the first time, the problem to design a revenue mechanism that ensures profits both to individual search engines and FSEs as a mechanism design problem. To this end, we extend the sponsored search auction models and we discuss possibility and impossibility results on the implementation of an incentive compatible mechanism. Specifically, we develop an execution-contingent VCG (where payments depend on the observed click behavior) that satisfies both individual rationality and weak budget balance in expectation.

【Keywords】:

98. Market Manipulation with Outside Incentives.

Paper Link】 【Pages】:

【Authors】: Yiling Chen ; Xi Alice Gao ; Rick Goldstein ; Ian A. Kash

【Abstract】: Much evidence has shown that prediction markets, when used in isolation, can effectively aggregate dispersed information about uncertain future events and produce remarkably accurate forecasts. However, if the market prediction will be used for decision making, a strategic participant with a vested interest in the decision outcome may want to manipulate the market prediction in order to influence the resulting decision. The presence of such incentives outside of the market would seem to damage information aggregation because of the potential distrust among market participants. While this is true under some conditions, we find that, if the existence of such incentives is certain and common knowledge, then in many cases, there exists a separating equilibrium for the market where information is fully aggregated. This equilibrium also maximizes social welfare for convex outside payoff functions. At this equilibrium, the participant with outside incentives makes a costly move to gain the trust of other participants. When the existence of outside incentives is uncertain, however, trust cannot be established between players if the outside incentive is sufficiently large and we lose the separability in equilibrium.

【Keywords】:

99. Parameterized Complexity of Problems in Coalitional Resource Games.

Paper Link】 【Pages】:

【Authors】: Rajesh Hemant Chitnis ; MohammadTaghi Hajiaghayi ; Vahid Liaghat

【Abstract】: Coalition formation is a key topic in multi-agent systems. Coalitions enable agents to achieve goals that they may nothave been able to achieve on their own. Previous work hasshown problems in coalition games to be computationally hard. Wooldridge and Dunne (Artifi. Intell. 2006) studied the classical computational complexity of several natural decision problems in Coalitional Resource Games (CRG) - games in which each agent is endowed with a set of resources and coalitions can bring about a set of goals if they are collectively endowed with the necessary amount of resources. The input of coalitional resource games bundles together several elements, e.g., the agent set Ag, the goal set G, the resource set R, etc. Shrot et al. (AAMAS 2009) examine coalition formation problems in the CRG model using the theory of Parameterized Complexity. Their refined analysis shows that not all parts of input act equal - some instances of the problem are indeed tractable while others still remain intractable.We answer an important question left open by Shrot, Aumann,and Kraus by showing that the SC Problem (checking whether a Coalition is Successful) is W[1]-hard when parameterized by the size of the coalition. Then via a single theme of reduction from SC, we are able to show that various problems related to resources, resource bounds, and resource conflicts introduced by Wooldridge et al. are (i) W[1]-hard or co-W[1]-hard w.r.t the size of the coalition; and (ii) Para-NP hard or co-Para-NP-hard w.r.t |R|. When parameterized by |G| or |R| + |Ag|, we give a general algorithm which proves that these problems are indeed tractable.

【Keywords】:

100. Optimal Envy-Free Cake Cutting.

Paper Link】 【Pages】:

【Authors】: Yuga J. Cohler ; John K. Lai ; David C. Parkes ; Ariel D. Procaccia

【Abstract】: We consider the problem of fairly dividing a heterogeneous divisible good among agents with different preferences. Previous work has shown that envy-free allocations, i.e., where each agent prefers its own allocation to any other, may not be efficient, in the sense of maximizing the total value of the agents. Our goal is to pinpoint the most efficient allocations among all envy-free allocations. We provide tractable algorithms for doing so under different assumptions regarding the preferences of the agents.

【Keywords】:

101. Commitment to Correlated Strategies.

Paper Link】 【Pages】:

【Authors】: Vincent Conitzer ; Dmytro Korzhyk

【Abstract】: The standard approach to computing an optimal mixed strategy to commit to is based on solving a set of linear programs, one for each of the follower's pure strategies. We show that these linear programs can be naturally merged into a single linear program; that this linear program can be interpreted as a formulation for the optimal correlated strategy to commit to, giving an easy proof of a result by von Stengel and Zamir that the leader's utility is at least the utility she gets in any correlated equilibrium of the simultaneous-move game; and that this linear program can be extended to compute optimal correlated strategies to commit to in games of three or more players. (Unlike in two-player games, in games of three or more players, the notions of optimal mixed and correlated strategies to commit to are truly distinct.) We give examples, and provide experimental results that indicate that for 50x50 games, this approach is usually significantly faster than the multiple-LPs approach.

【Keywords】:

102. Dominating Manipulations in Voting with Partial Information.

Paper Link】 【Pages】:

【Authors】: Vincent Conitzer ; Toby Walsh ; Lirong Xia

【Abstract】: We consider manipulation problems when the manipulator only has partial information about the votes of the non-manipulators. Such partial information is described by an {\em information set}, which is the set of profiles of the non-manipulators that are indistinguishable to the manipulator. Given such an information set, a {\em dominating manipulation} is a non-truthful vote that the manipulator can cast which makes the winner at least as preferable (and sometimes more preferable) as the winner when the manipulator votes truthfully. When the manipulator has full information, computing whether or not there exists a dominating manipulation is in P for many common voting rules (by known results). We show that when the manipulator has no information, there is no dominating manipulation for many common voting rules. When the manipulator's information is represented by partial orders and only a small portion of the preferences are unknown, computing a dominating manipulation is NP-hard for many common voting rules. Our results thus throw light on whether we can prevent strategic behavior by limiting information about the votes of other voters.

【Keywords】:

103. On Expressing Value Externalities in Position Auctions.

Paper Link】 【Pages】:

【Authors】: Florin Constantin ; Malvika Rao ; Chien-Chung Huang ; David C. Parkes

【Abstract】: We introduce a bidding language for expressing negative value externalities in position auctions for online advertising. The unit-bidder constraints (UBC) language allows a bidder to condition a bid on its allocated slot and on the slots allocated to other bidders. We introduce a natural extension of the Generalized Second Price (GSP) auction, the expressive GSP (eGSP) auction, that induces truthful revelation of constraints for a rich subclass of unit-bidder types, namely downward-monotonic UBC. We establish the existence of envy-free Nash equilibrium in eGSP under a further restriction to a subclass of exclusion constraints, for which the standard GSP has no pure strategy Nash equilibrium. The equilibrium results are obtained by reduction to equilibrium analysis for reserve price GSP (Even-Dar et al. 2008). In considering the winner determination problem, which is NP-hard, we bound the approximation ratio for social welfare in eGSP and provide parameterized complexity results.

【Keywords】:

104. Learning in Repeated Games with Minimal Information: The Effects of Learning Bias.

Paper Link】 【Pages】:

【Authors】: Jacob W. Crandall ; Asad Ahmed ; Michael A. Goodrich

【Abstract】: Automated agents for electricity markets, social networks, and other distributed networks must repeatedly interact with other intelligent agents, often without observing associates' actions or payoffs (i.e., minimal information). Given this reality, our goal is to create algorithms that learn effectively in repeated games played with minimal information. As in other applications of machine learning, the success of a learning algorithm in repeated games depends on its learning bias. To better understand what learning biases are most successful, we analyze the learning biases of previously published multi-agent learning (MAL) algorithms. We then describe a new algorithm that adapts a successful learning bias from the literature to minimal information environments. Finally, we compare the performance of this algorithm with ten other algorithms in repeated games played with minimal information.

【Keywords】:

105. Complexity of and Algorithms for Borda Manipulation.

Paper Link】 【Pages】:

【Authors】: Jessica Davies ; George Katsirelos ; Nina Narodytska ; Toby Walsh

【Abstract】: We prove that it is NP-hard for a coalition of two manipulators to compute how to manipulate the Borda voting rule. This resolves one of the last open problems in the computational complexity of manipulating common voting rules. Because of this NP-hardness, we treat computing a manipulation as an approximation problem where we try to minimize the number of manipulators. Based on ideas from bin packing and multiprocessor scheduling, we propose two new approximation methods to compute manipulations of the Borda rule. Experiments show that these methods significantly outperform the previous best known approximation method. We are able to find optimal manipulations in almost all the randomly generated elections tested. Our results suggest that, whilst computing a manipulation of the Borda rule by a coalition is NP-hard, computational complexity may provide only a weak barrier against manipulation in practice.

【Keywords】:

106. A Functional Analysis of Historical Memory Retrieval Bias in the Word Sense Disambiguation Task.

Paper Link】 【Pages】:

【Authors】: Nate Derbinsky ; John E. Laird

【Abstract】: Effective access to knowledge within large declarative memory stores is one challenge in the development and understanding of long-living, generally intelligent agents. We focus on a sub-component of this problem: given a large store of knowledge, how should an agent's task-independent memory mechanism respond to an ambiguous cue, one that pertains to multiple previously encoded memories. A large body of cognitive modeling work suggests that human memory retrievals are biased in part by the recency and frequency of past memory access. In this paper, we evaluate the functional benefit of a set of memory retrieval heuristics that incorporate these biases, in the context of the word sense disambiguation task, in which an agent must identify the most appropriate word meaning in response to an ambiguous linguistic cue. In addition, we develop methods to integrate these retrieval biases within a task-independent declarative memory system implemented in the Soar cognitive architecture and evaluate their effectiveness and efficiency in three commonly used semantic concordances.

【Keywords】:

107. Computing an Extensive-Form Perfect Equilibrium in Two-Player Games.

Paper Link】 【Pages】:

【Authors】: Nicola Gatti ; Claudio Iuliano

【Abstract】: Equilibrium computation in games is currently considered one of the most challenging issues in AI. In this paper, we provide, to the best of our knowledge, the first algorithm to compute a Selten's extensive-form perfect equilibrium (EFPE) with two--player games. EFPE refines the Nash equilibrium requiring the equilibrium to be robust to slight perturbations of both players' behavioral strategies. Our result puts the computation of an EFPE into the PPAD class, leaving open the question whether or not the problem is hard. Finally, we experimentally evaluate the computational time spent to find an EFPE and some relaxations of EFPE.

【Keywords】:

108. VCG Redistribution with Gross Substitutes.

Paper Link】 【Pages】:

【Authors】: Mingyu Guo

【Abstract】: For the problem of allocating resources among multiple strategic agents, the well-known Vickrey-Clarke-Groves (VCG) mechanism is efficient, strategy-proof, and it never incurs a deficit. However, in general, under the VCG mechanism, payments flow out of the system of agents, which reduces the agents' utilities. VCG redistribution mechanisms aim to return as much of the VCG payments as possible back to the agents, without affecting the desirable properties of the VCG mechanism. Most previous research on VCG redistribution mechanisms has focused on settings with homogeneous items and/or settings with unit-demand agents. In this paper, we study VCG redistribution mechanisms in the more general setting of combinatorial auctions. We show that when the gross substitutes condition holds, we are able to design mechanisms that guarantee to redistribute a large fraction of the VCG payments.

【Keywords】:

109. Automated Action Abstraction of Imperfect Information Extensive-Form Games.

Paper Link】 【Pages】:

【Authors】: John Alexander Hawkin ; Robert Holte ; Duane Szafron

【Abstract】: Multi-agent decision problems can often be formulated as extensive-form games. We focus on imperfect information extensive-form games in which one or more actions at many decision points have an associated continuous or many-valued parameter. A stock trading agent, in addition to deciding whether to buy or not, must decide how much to buy. In no-limit poker, in addition to selecting a probability for each action, the agent must decide how much to bet for each betting action. Selecting values for these parameters makes these games extremely large. Two-player no-limit Texas Hold'em poker with stacks of 500 big blinds has approximately 10 71 states, which is more than 10 50 times more states than two-player limit Texas Hold'em. The main contribution of this paper is a technique that abstracts a game's action space by selecting one, or a small number, of the many values for each parameter. We show that strategies computed using this new algorithm for no-limit Leduc poker exhibit significant utility gains over epsilon-Nash equilibrium strategies computed with standard, hand-crafted parameter value abstractions.

【Keywords】:

110. A Game-Theoretic Approach to Influence in Networks.

Paper Link】 【Pages】:

【Authors】: Mohammad Tanvir Irfan ; Luis E. Ortiz

【Abstract】: We propose influence games, a new class of graphical games, as a model of the behavior of large but finite networked populations. Grounded in non-cooperative game theory, we introduce a new approach to the study of influence in networks that captures the strategic aspects of complex interactions in the network. We study computational problems on influence games, including the identification of the most influential nodes. We characterize the computational complexity of various problems in influence games, propose several heuristics for the hard cases, and design approximation algorithms, with provable guarantees, for the most influential nodes problem.

【Keywords】:

111. A Kernel-Based Iterative Combinatorial Auction.

Paper Link】 【Pages】:

【Authors】: Sébastien Lahaie

【Abstract】: This paper describes an iterative combinatorial auction for single-minded bidders that offers modularity in the choice of price structure, drawing on ideas from kernel methods and the primal-dual paradigm of auction design. In our implementation, the auction is able to automatically detect, as the rounds progress, whether price expressiveness must be increased to clear the market. The auction also features a configurable step size which can be tuned to trade-off between monotonicity in prices and the number of bidding rounds, with no impact on efficiency. An empirical evaluation against a state of the art ascending-price auction demonstrates the performance gains that can be obtained in efficiency, revenue, and rounds to convergence through various configurations of our design.

【Keywords】:

112. A Distributed Anytime Algorithm for Dynamic Task Allocation in Multi-Agent Systems.

Paper Link】 【Pages】:

【Authors】: Kathryn Sarah Macarthur ; Ruben Stranders ; Sarvapali D. Ramchurn ; Nicholas R. Jennings

【Abstract】: We introduce a novel distributed algorithm for multi-agent task allocation problems where the sets of tasks and agents constantly change over time. We build on an existing anytime algorithm (fast-max-sum), and give it significant new capa- bilities: namely, an online pruning procedure that simplifies the problem, and a branch-and-bound technique that reduces the search space. This allows us to scale to problems with hundreds of tasks and agents. We empirically evaluate our algorithm against established benchmarks and find that, even in such large environments, a solution is found up to 31% faster, and with up to 23% more utility, than state-of-the-art approximation algorithms. In addition, our algorithm sends up to 30% fewer messages than current approaches when the set of agents or tasks changes.

【Keywords】:

113. Quick Polytope Approximation of All Correlated Equilibria in Stochastic Games.

Paper Link】 【Pages】:

【Authors】: Liam MacDermed ; Karthik Sankaran Narayan ; Charles Lee Isbell Jr. ; Lora Weiss

【Abstract】: Stochastic or Markov games serve as reasonable models for a variety of domains from biology to computer security, and are appealing due to their versatility. In this paper we address the problem of finding the complete set of correlated equilibria for general-sum stochastic games with perfect information. We present QPACE — an algorithm orders of magnitude more efficient than previous approaches while maintaining a guarantee of convergence and bounded error. Finally, we validate our claims and demonstrate the limits of our algorithm with extensive empirical tests.

【Keywords】:

114. Manipulation of Nanson's and Baldwin's Rules.

Paper Link】 【Pages】:

【Authors】: Nina Narodytska ; Toby Walsh ; Lirong Xia

【Abstract】: Nanson's and Baldwin's voting rules selecta winner by successively eliminatingcandidates with low Borda scores. We showthat these rules have a number of desirablecomputational properties. In particular,with unweighted votes, it isNP-hard to manipulate either rule with one manipulator, whilstwith weighted votes, it isNP-hard to manipulate either rule with a small number ofcandidates and a coalition of manipulators.As only a couple of other voting rulesare known to be NP-hard to manipulatewith a single manipulator, Nanson'sand Baldwin's rules appearto be particularly resistant to manipulation from a theoretical perspective.We also propose a number of approximation methodsfor manipulating these two rules.Experiments demonstrate that both rules areoften difficult to manipulate in practice.These results suggest that elimination stylevoting rules deserve further study.

【Keywords】:

115. Constrained Coalition Formation.

Paper Link】 【Pages】:

【Authors】: Talal Rahwan ; Tomasz P. Michalak ; Edith Elkind ; Piotr Faliszewski ; Jacek Sroka ; Michael Wooldridge ; Nicholas R. Jennings

【Abstract】: The conventional model of coalition formation considers every possible subset of agents as a potential coalition. However, in many real-world applications, there are inherent constraints on feasible coalitions: for instance, certain agents may be prohibited from being in the same coalition, or the coalition structure may be required to consist of coalitions of the same size. In this paper, we present the first systematic study of constrained coalition formation (CCF). We propose a general framework for this problem, and identify an important class of CCF settings, where the constraints specify which groups of agents should/should not work together. We describe a procedure that transforms such constraints into a structured input that allows coalition formation algorithms to identify, without any redundant computations, all the feasible coalitions. We then use this procedure to develop an algorithm for generating an optimal (welfare-maximizing) constrained coalition structure, and show that it outperforms existing state-of-the-art approaches by several orders of magnitude.

【Keywords】:

116. Campaign Management under Approval-Driven Voting Rules.

Paper Link】 【Pages】:

【Authors】: Ildikó Schlotter ; Piotr Faliszewski ; Edith Elkind

【Abstract】: Approval-like voting rules, such as Sincere-Strategy Preference-Based Approval voting (SP-AV), the Bucklin rule (an adaptive variant of k-Approval voting), and the Fallback rule (an adaptive variant of SP-AV) have many desirable properties: for example, they are easy to understand and encourage the candidates to choose electoral platforms that have a broad appeal. In this paper, we investigate both classic and parameterized computational complexity of electoral campaign management under such rules. We focus on two methods that can be used to promote a given candidate: asking voters to move this candidate upwards in their preference order or asking them to change the number of candidates they approve of. We show that finding an optimal campaign management strategy of the first type is easy for both Bucklin and Fallback. In contrast, the second method is computationally hard even if the degree to which we need to affect the votes is small. Nevertheless, we identify a large class of scenarios that admit a fixed-parameter tractable algorithm.

【Keywords】:

117. M-Unit EigenAnt: An Ant Algorithm to Find the M Best Solutions.

Paper Link】 【Pages】:

【Authors】: Sameena Shah ; Jayadeva ; Ravi Kothari ; Suresh Chandra

【Abstract】: In this paper, we shed light on how powerful congestion control based on local interactions may be obtained. We show how ants can use repellent pheromones and incorporate the effect of crowding to avoid traffic congestion on the optimal path. Based on these interactions, we propose an ant algorithm, the M-unit EigenAnt algorithm, that leads to the selection of the M shortest paths. The ratio of selection of each of these paths is also optimal and regulated by an optimal amount of pheromone on each of them. To the best of our knowledge, the M -unit EigenAnt algorithm is the first antalgorithm that explicitly ensures the selection of the M shortest paths and regulates the amount of pheromone on them such that it is asymptotically optimal. In fact, it is in contrast with most ant algorithms that aim to discover just a single best path. We provide its convergence analysis and show that the steady state distribution of pheromone aligns with the eigenvectors of the cost matrix, and thus is related to its measure of quality. We also provide analysis to show that this property ensues even when the food is moved or path lengths change during foraging. We show that this behavior is robust in the presence of fluctuations and quickly reflects the change in the M optimal solutions. This makes it suitable for not only distributed applications butalso dynamic ones as well. Finally, we provide simulation results for the convergence to the optimal solution under different initial biases, dynamism in lengths of paths, and discovery of new paths.

【Keywords】:

118. Efficiency and Privacy Tradeoffs in Mechanism Design.

Paper Link】 【Pages】:

【Authors】: Xin Sui ; Craig Boutilier

【Abstract】: A key problem in mechanism design is the construction of protocols that reach socially efficient decisions with minimal information revelation. This can reduce agent communication, and further, potentially increase privacy in the sense that agents reveal no more private information than is needed to determine an optimal outcome. This is not always possible: previous work has explored the tradeoff between communication cost and efficiency, and more recently, communication and privacy. We explore a third dimension: the tradeoff between privacy and efficiency. By sacrificing efficiency, we can improve the privacy of a variety of existing mechanisms. We analyze these tradeoffs in both second-price auctions and facility location problems (introducing new incremental mechanisms for facility location along the way). Our results show that sacrifices in efficiency can provide gains in privacy (and communication), in both the average and worst case.

【Keywords】:

119. Dominant-Strategy Auction Design for Agents with Uncertain, Private Values.

Paper Link】 【Pages】:

【Authors】: David Robert Martin Thompson ; Kevin Leyton-Brown

【Abstract】: We study the problem of designing auctions for agents who incur a cost if they choose to learn about their own preferences. We reformulate the revelation principle for use with such deliberative agents. Then we characterize the set of single-good auctions giving rise to dominant strategies for deliberative agents whose values are independent and private. Interestingly, this set of dominant-strategy mechanisms is exactly the set of sequential posted-price auctions, a class of mechanisms that has received much recent attention.  

【Keywords】:

120. Incentive-Compatible Escrow Mechanisms.

Paper Link】 【Pages】:

【Authors】: Jens Witkowski ; Sven Seuken ; David C. Parkes

【Abstract】: The most prominent way to establish trust between buyers and sellers on online auction sites are reputation mechanisms. Two drawbacks of this approach are the reliance on the seller being long-lived and the susceptibility to whitewashing. In this paper, we introduce so-called escrow mechanisms that avoid these problems by installing a trusted intermediary which forwards the payment to the seller only if the buyer acknowledges that the good arrived in the promised condition. We address the incentive issues that arise and design an escrow mechanism that is incentive-compatible, efficient, interim individually rational and ex ante budget-balanced. In contrast to previous work on trust and reputation, our approach does not rely on knowing the sellers' cost functions or the distribution of buyer valuations.

【Keywords】:

121. Risk-Averse Strategies for Security Games with Execution and Observational Uncertainty.

Paper Link】 【Pages】:

【Authors】: Zhengyu Yin ; Manish Jain ; Milind Tambe ; Fernando Ordóñez

【Abstract】: Attacker-defender Stackelberg games have become a popular game-theoretic approach for security with deployments for LAX Police, the FAMS and the TSA. Unfortunately, most of the existing solution approaches do not model two key uncertainties of the real-world: there may be noise in the defender's execution of the suggested mixed strategy and/or the observations made by an attacker can be noisy. In this paper, we provide a framework to model these uncertainties, and demonstrate that previous strategies perform poorly in such uncertain settings. We also provide RECON, a novel algorithm that computes strategies for the defender that are robust to such uncertainties, and provide heuristics that further improve RECON's efficiency.

【Keywords】:

122. Coordinated Multi-Agent Reinforcement Learning in Networked Distributed POMDPs.

Paper Link】 【Pages】:

【Authors】: Chongjie Zhang ; Victor R. Lesser

【Abstract】: In many multi-agent applications such as distributed sensor nets, a network of agents act collaboratively under uncertainty and local interactions. Networked Distributed POMDP (ND-POMDP) provides a framework to model such cooperative multi-agent decision making. Existing work on ND-POMDPs has focused on offline techniques that require accurate models, which are usually costly to obtain in practice. This paper presents a model-free, scalable learning approach that synthesizes multi-agent reinforcement learning (MARL) and distributed constraint optimization (DCOP). By exploiting structured interaction in ND-POMDPs, our approach distributes the learning of the joint policy and employs DCOP techniques to coordinate distributed learning to ensure the global learning performance. Our approach can learn a globally optimal policy for ND-POMDPs with a property called groupwise observability. Experimental results show that, with communication during learning and execution, our approach significantly outperforms the nearly-optimal non-communication policies computed offline.

【Keywords】:

Multidisciplinary Topics 14

123. The Influence of Emotion Expression on Perceptions of Trustworthiness in Negotiation.

Paper Link】 【Pages】:

【Authors】: Dimitrios Antos ; Celso de Melo ; Jonathan Gratch ; Barbara J. Grosz

【Abstract】: When interacting with computer agents, people make inferences about various characteristics of these agents, such as their reliability and trustworthiness. These perceptions are significant, as they influence people's behavior towards the agents, and may foster or inhibit repeated interactions between them. In this paper we investigate whether computer agents can use the expression of emotion to influence human perceptions of trustworthiness. In particular, we study human-computer interactions within the context of a negotiation game, in which players make alternating offers to decide on how to divide a set of resources. A series of negotiation games between a human and several agents is then followed by a "trust game." In this game people have to choose one among several agents to interact with, as well as how much of their resources they will trust to it. Our results indicate that, among those agents that displayed emotion, those whose expression was in accord with their actions (strategy) during the negotiation game were generally preferred as partners in the trust game over those whose emotion expressions and actions did not mesh. Moreover, we observed that when emotion does not carry useful new information, it fails to strongly influence human decision-making behavior in a negotiation setting.

【Keywords】:

124. Co-Evolution of Selection and Influence in Social Networks.

Paper Link】 【Pages】:

【Authors】: Yoon-Sik Cho ; Greg Ver Steeg ; Aram Galstyan

【Abstract】: Many networks are complex dynamical systems, where both attributes of nodes and topology of the network (link structure) can change with time. We propose a model of co-evolving networks where both node attributes and network structure evolve under mutual influence. Specifically, we consider a mixed membership stochastic blockmodel, where the probability of observing a link between two nodes depends on their current membership vectors, while those membership vectors themselves evolve in the presence of a link between the nodes. Thus, the network is shaped by the interaction of stochastic processes describing the nodes, while the processes themselves are influenced by the changing network structure. We derive an efficient variational inference procedure for our model, and validate the model on both synthetic and real-world data.

【Keywords】:

125. Finding Answers and Generating Explanations for Complex Biomedical Queries.

Paper Link】 【Pages】:

【Authors】: Esra Erdem ; Yelda Erdem ; Halit Erdogan ; Umut Öztok

【Abstract】: We present new methods to efficiently answer complex queries overbiomedical ontologies and databases considering the relevant partsof these knowledge resources, and to generate shortest explanationsto justify these answers. Both algorithms rely on the high-levelrepresentation and efficient solvers of Answer Set Programming. Weapply these algorithms to find answers and explanations to some complexqueries related to drug discovery, over PharmGKB, DrugBank, BioGrid, CTD and Sider.

【Keywords】:

126. First-Order Logic with Counting for General Game Playing.

Paper Link】 【Pages】:

【Authors】: Lukasz Kaiser ; Lukasz Stafiniak

【Abstract】: General Game Players (GGPs) are programs which can play an arbitrary game given only its rules and the Game Description Language (GDL) is a variant of Datalog used in GGP competitions to specify the rules. GDL inherits from Datalog the use of Horn clauses as rules and recursion, but it too requires stratification and does not allow to use quantifiers. We present an alternative formalism for game description which is based on first-order logic (FO). States of the game are represented by relational structures, legal moves by structure rewriting rules guarded by FO formulas, and the goals of the players by formulas which extend FO with counting. The advantage of our formalism comes from more explicit state representationcand from the use of quantifiers in formulas. We show how to exploit existential quantification in players' goals to generate heuristics for evaluating positions in the game. The derived heuristics are good enough for a basic alpha-beta agent to win against state of the art GGP.

【Keywords】:

127. Grammatical Error Detection for Corrective Feedback Provision in Oral Conversations.

Paper Link】 【Pages】:

【Authors】: Sungjin Lee ; Hyungjong Noh ; Kyusong Lee ; Gary Geunbae Lee

【Abstract】: The demand for computer-assisted language learning systems that can provide corrective feedback on language learners’ speaking has increased. However, it is not a trivial task to detect grammatical errors in oral conversations because of the unavoidable errors of automatic speech recognition systems. To provide corrective feedback, a novel method to detect grammatical errors in speaking performance is proposed. The proposed method consists of two sub-models: the grammaticality-checking model and the error-type classification model. We automatically generate grammatical errors that learners are likely to commit and construct error patterns based on the articulated errors. When a particular speech pattern is recognized, the grammaticality-checking model performs a binary classification based on the similarity between the error patterns and the recognition result using the confidence score. The error-type classification model chooses the error type based on the most similar error pattern and the error frequency extracted from a learner corpus. The grammaticality checking method largely outperformed the two comparative models by 56.36% and 42.61% in F-score while keeping the false positive rate very low. The error-type classification model exhibited very high performance with a 99.6% accuracy rate. Because high precision and a low false positive rate are important criteria for the language-tutoring setting, the proposed method will be helpful for intelligent computer-assisted language learning systems.

【Keywords】:

128. Social Relations Model for Collaborative Filtering.

Paper Link】 【Pages】:

【Authors】: Wu-Jun Li ; Dit-Yan Yeung

【Abstract】: We propose a novel probabilistic model for collaborative filtering (CF), called SRMCoFi, which seamlessly integrates both linear and bilinear random effects into a principled framework. The formulation of SRMCoFi is supported by both social psychological experiments and statistical theories. Not only can many existing CF methods be seen as special cases of SRMCoFi, but it also integrates their advantages while simultaneously overcoming their disadvantages. The solid theoretical foundation of SRMCoFi is further supported by promising empirical results obtained in extensive experiments using real CF data sets on movie ratings.

【Keywords】:

129. Comparing Agents' Success against People in Security Domains.

Paper Link】 【Pages】:

【Authors】: Raz Lin ; Sarit Kraus ; Noa Agmon ; Samuel Barrett ; Peter Stone

【Abstract】: The interaction of people with autonomous agents has become increasingly prevalent. Some of these settings include security domains, where people can be characterized as uncooperative, hostile, manipulative, and tending to take advantage of the situation for their own needs. This makes it challenging to design proficient agents to interact with people in such environments. Evaluating the success of the agents automatically before evaluating them with people or deploying them could alleviate this challenge and result in better designed agents. In this paper we show how Peer Designed Agents (PDAs) -- computer agents developed by human subjects -- can be used as a method for evaluating autonomous agents in security domains. Such evaluation can reduce the effort and costs involved in evaluating autonomous agents interacting with people to validate their efficacy. Our experiments included more than 70 human subjects and 40 PDAs developed by students. The study provides empirical support that PDAs can be used to compare the proficiency of autonomous agents when matched with people in security domains.

【Keywords】:

130. Bayesian Learning of Generalized Board Positions for Improved Move Prediction in Computer Go.

Paper Link】 【Pages】:

【Authors】: Martin Michalowski ; Mark Boddy ; Mike Neilsen

【Abstract】: Computer Go presents a challenging problem for machine learning agents. With the number of possible board states estimated to be larger than the number of hydrogen atoms in the universe, learning effective policies or board evaluation functions is extremely difficult. In this paper we describe Cortigo, a system that efficiently and autonomously learns useful generalizations for large state-space classification problems such as Go. Cortigo uses a hierarchical generative model loosely related to the human visual cortex to recognize Go board positions well enough to suggest promising next moves. We begin by briefly describing and providing motivation for research in the computer Go domain. We describe Cortigo’s ability to learn predictive models based on large subsets of the Go board and demonstrate how using Cortigo’s learned models as additive knowledge in a state-of-the-art computer Go player (Fuego) significantly improves its playing strength.

【Keywords】:

131. Composite Social Network for Predicting Mobile Apps Installation.

Paper Link】 【Pages】:

【Authors】: Wei Pan ; Nadav Aharony ; Alex Pentland

【Abstract】: We have carefully instrumented a large portion of the population living in a university graduate dormitory by giving participants Android smart phones running our sensing software. In this paper, we propose the novel problem of predicting mobile application (known as “apps”) installation using social networks and explain its challenge. Modern smart phones, like the ones used in our study, are able to collect different social networks using built-in sensors. (e.g. Bluetooth proximity network, call log network, etc) While this information is accessible to app market makers such as the iPhone AppStore, it has not yet been studied how app market makers can use these information for marketing research and strategy development. We develop a simple computational model to better predict app installation by using a composite network computed from the different networks sensed by phones. Our model also captures individual variance and exogenous factors in app adoption. We show the importance of considering all these factors in predicting app installations, and we observe the surprising result that app installation is indeed predictable. We also show that our model achieves the best results compared with generic approaches.

【Keywords】:

132. Human Spatial Relational Reasoning: Processing Demands, Representations, and Cognitive Model.

Paper Link】 【Pages】:

【Authors】: Marco Ragni ; Sven Brüssow

【Abstract】: Empirical findings indicate that humans draw infer- ences about spatial arrangements by constructing and manipulating mental models which are internal representations of objects and relations in spatial working memory. Central to the Mental Model Theory (MMT), is the assumption that the human reasoning process can be divided into three phases: (i) Mental model construction, (ii) model inspection, and (iii) model validation. The MMT can be formalized with respect to a computational model, connecting the reasoning process to operations on mental model representations. In this respect a computational model has been implemented in the cognitive architecture ACT-R capable of explaining human reasoning difficulty by the number of model operations. The presented ACT-R model allows simulation of psychological findings about spatial reasoning problems from a previous study that investigated conventional behavioral data such as response times and error rates in the context of certain mental model construction principles.

【Keywords】:

133. Intrinsic Chess Ratings.

Paper Link】 【Pages】:

【Authors】: Kenneth Wingate Regan ; Guy McCrossan Haworth

【Abstract】: This paper develops and tests formulas for representing playing strength at chess by the quality of moves played, rather than by the results of games. Intrinsic quality is estimated via evaluations given by computer chess programs run to high depth, ideally so that their playing strength is sufficiently far ahead of the best human players as to be a relatively omniscient' guide. Several formulas, each having intrinsic skill parameters s forsensitivity' and c for consistency', are argued theoretically and tested by regression on large sets of tournament games played by humans of varying strength as measured by the internationally standard Elo rating system. This establishes a correspondence between Elo rating and the parameters. A smooth correspondence is shown between statistical results and the century points on the Elo scale, and ratings are shown to have stayed quite constant over time. That is, there has been little or norating inflation'. The theory and empirical results are transferable to other rational-choice settings in which the alternatives have well-defined utilities, but in which complexity and bounded information constrain the perception of the utility values.

【Keywords】:

134. The Epistemic Logic Behind the Game Description Language.

Paper Link】 【Pages】:

【Authors】: Ji Ruan ; Michael Thielscher

【Abstract】: A general game player automatically learns to play arbitrary new games solely by being told their rules. For this purpose games are specified in the game description language GDL, a variant of Datalog with function symbols and a few known keywords. In its latest version GDL allows to describe nondeterministic games with any number of players who may have imperfect, asymmetric information. We analyse the epistemic structure and expressiveness of this language in terms of epistemic modal logic and present two main results: The operational semantics of GDL entails that the situation at any stage of a game can be characterised by a multi-agent epistemic (i.e., S5-) model; (2) GDL is sufficiently expressive to model any situation that can be described by a (finite) multi-agent epistemic model.

【Keywords】:

135. Reasoning About General Games Described in GDL-II.

Paper Link】 【Pages】:

【Authors】: Stephan Schiffel ; Michael Thielscher

【Abstract】: Recently the general Game Description Language (GDL) has been extended so as to cover arbitrary games with incomplete/imperfect information. Learning — without human intervention — to play such games poses a reasoning challenge for general game-playing systems that is much more intricate than in case of complete information games. Action formalisms like the Situation Calculus have been developed for precisely this purpose. In this paper we present a full embedding of the Game Description Language into the Situation Calculus (with Scherl and Levesque's knowledge fluent ). We formally prove that this provides a sound and complete reasoning method for players' knowledge about game states as well as about the knowledge of the other players.

【Keywords】:

136. Co-Training as a Human Collaboration Policy.

Paper Link】 【Pages】:

【Authors】: Xiaojin Zhu ; Bryan R. Gibson ; Timothy T. Rogers

【Abstract】: We consider the task of human collaborative category learning, where two people work together to classify test items into appropriate categories based on what they learn from a training set. We propose a novel collaboration policy based on the Co-Training algorithm in machine learning, in which the two people play the role of the base learners. The policy restricts each learner's view of the data and limits their communication to only the exchange of their labelings on test items. In a series of empirical studies, we show that the Co-Training policy leads collaborators to jointly produce unique and potentially valuable classification outcomes that are not generated under other collaboration policies. We further demonstrate that these observations can be explained with appropriate machine learning models.

【Keywords】:

Natural Language Processing 13

Paper Link】 【Pages】:

【Authors】: David L. Chen ; Raymond J. Mooney

【Abstract】: The ability to understand natural-language instructions is critical to building intelligent agents that interact with humans. We present a system that learns to transform natural-language navigation instructions into executable formal plans. Given no prior linguistic knowledge, the system learns by simply observing how humans follow navigation instructions. The system is evaluated in three complex virtual indoor environments with numerous objects and landmarks. A previously collected realistic corpus of complex English navigation instructions for these environments is used for training and testing data. By using a learned lexicon to refine inferred plans and a supervised learner to induce a semantic parser, the system is able to automatically learnto correctly interpret a reasonable fraction of the complex instructions in this corpus.

【Keywords】:

138. A Simple and Effective Unsupervised Word Segmentation Approach.

Paper Link】 【Pages】:

【Authors】: Songjian Chen ; Yabo Xu ; Huiyou Chang

【Abstract】: In this paper, we propose a new unsupervised approach for word segmentation. The core idea of our approach is a novel word induction criterion called WordRank, which estimates the goodness of word hypotheses (character or phoneme sequences). We devise a method to derive exterior word boundary information from the link structures of adjacent word hypotheses and incorporate interior word boundary information to complete the model. In light of WordRank, word segmentation can be modeled as an optimization problem. A Viterbi-styled algorithm is developed for the search of the optimal segmentation. Extensive experiments conducted on phonetic transcripts as well as standard Chinese and Japanese data sets demonstrate the effectiveness of our approach. On the standard Brent version of Bernstein-Ratner corpora, our approach outperforms the state-of-the-art Bayesian models by more than 3%. Plus, our approach is simpler and more efficient than the Bayesian methods. Consequently, our approach is more suitable for real-world applications.

【Keywords】:

Paper Link】 【Pages】:

【Authors】: Jennifer Chu-Carroll ; James Fan

【Abstract】: Most existing Question Answering (QA) systems adopt a type-and-generate approach to candidate generation that relies on a pre-defined domain ontology. This paper describes a type independent search and candidate generation paradigm for QA that leverages Wikipedia characteristics. This approach is particularly useful for adapting QA systems to domains where reliable answer type identification and type-based answer extraction are not available. We present a three-pronged search approach motivated by relations an answer-justifying title-oriented document may have with the question/answer pair. We further show how Wikipedia metadata such as anchor texts and redirects can be utilized to effectively extract candidate answers from search results without a type ontology. Our experimental results show that our strategies obtained high binary recall in both search and candidate generation on TREC questions, a domain that has mature answer type extraction technology, as well as on Jeopardy! questions, a domain without such technology. Our high-recall search and candidate generation approach has also led to high overall QA performance in Watson, our end-to-end system.

【Keywords】:

140. Lossy Conservative Update (LCU) Sketch: Succinct Approximate Count Storage.

Paper Link】 【Pages】:

【Authors】: Amit Goyal ; Hal Daumé III

【Abstract】: In this paper, we propose a variant of the conservativeupdate Count-Min sketch to further reduce the overestimation error incurred. Inspired by ideas from lossy counting, we divide a stream of items into multiple windows, and decrement certain counts in the sketch at window boundaries. We refer to this approach as a lossy conservative update (LCU). The reduction in overestimation error of counts comes at the cost of introducing under-estimation error in counts. However, in our intrinsic evaluations, we show that the reduction in overestimation is much greater than the under-estimation error introduced by our method LCU. We apply our LCU framework to scale distributional similarity computations to web-scale corpora. We show that this technique is more efficient in terms of memory, and time, and more robust than conservative update with Count-Min (CU) sketch on this task.

【Keywords】:

141. Semantic Relatedness Using Salient Semantic Analysis.

Paper Link】 【Pages】:

【Authors】: Samer Hassan ; Rada Mihalcea

【Abstract】: This paper introduces a novel method for measuring semantic relatedness using semantic profiles constructed from salient encyclopedic features. The model is built on the notion that the meaning of a word can be characterized by the salient concepts found in its immediate context. In addition to being computationally efficient, the new model has superior performance and remarkable consistency when compared to both knowledge-based and corpus-based state-of-the-art semantic relatedness models.

【Keywords】:

142. Partially Supervised Text Classification with Multi-Level Examples.

Paper Link】 【Pages】:

【Authors】: Tao Liu ; Xiaoyong Du ; Yong-Dong Xu ; Minghui Li ; Xiaolong Wang

【Abstract】: Partially supervised text classification has received great research attention since it only uses positive and unlabeled examples as training data. This problem can be solved by automatically labeling some negative (and more positive) examples from unlabeled examples before training a text classifier. But it is difficult to guarantee both high quality and quantity of the new labeled examples. In this paper, a multi-level example based learning method for partially supervised text classification is proposed, which can make full use of all unlabeled examples. A heuristic method is proposed to assign possible labels to unlabeled examples and partition them into multiple levels according to their labeling confidence. A text classifier is trained on these multi-level examples using weighted support vector machines. Experiments show that the multi-level example based learning method is effective for partially supervised text classification, and outperforms the existing popular methods such as Biased-SVM, ROC-SVM, S-EM and WL.

【Keywords】:

143. Enhancing Semantic Role Labeling for Tweets Using Self-Training.

Paper Link】 【Pages】:

【Authors】: Xiaohua Liu ; Kuan Li ; Ming Zhou ; Zhongyang Xiong

【Abstract】: Semantic Role Labeling (SRL) for tweets is a meaningful task that can benefit a wide range of applications such as fine-grained information extraction and retrieval from tweets. One main challenge of the task is the lack of annotated tweets, which is required to train a statistical model. We introduce self-training to SRL, leveraging abundant unlabeled tweets to alleviate its depending on annotated tweets. A novel strategy of tweet selection is presented, ensuring the chosen tweets are both correct and informative. More specifically, the correctness is estimated according to the labeling confidences and agreement of two Conditional Random Fields based labelers, which are trained on the randomly evenly spitted labeled data; while the informativeness is in proportion to the maximum distance between the tweet and the already selected tweets. We evaluate our method on a human annotated data set and show that bootstrapping improve a baseline by 3.4% F1.

【Keywords】:

144. Using Semantic Cues to Learn Syntax.

Paper Link】 【Pages】:

【Authors】: Tahira Naseem ; Regina Barzilay

【Abstract】: We present a method for dependency grammar induction that utilizes sparse annotations of semantic relations. This induction set-up is attractive because such annotations provide useful clues about the underlying syntactic structure, and they are readily available in many domains (e.g., info-boxes and HTML markup). Our method is based on the intuition that syntactic realizations of the same semantic predicate exhibit some degree of consistency. We incorporate this intuition in a directed graphical model that tightly links the syntactic and semantic structures. This design enables us to exploit syntactic regularities while still allowing for variations. Another strength of the model lies in its ability to capture non-local dependency relations. Our results demonstrate that even a small amount of semantic annotations greatly improves the accuracy of learned dependencies when tested on both in-domain and out-of-domain texts.

【Keywords】:

145. Exploiting Phase Transition in Latent Networks for Clustering.

Paper Link】 【Pages】:

【Authors】: Vahed Qazvinian ; Dragomir R. Radev

【Abstract】: In this paper, we model the pair-wise similarities of a setof documents as a weighted network with a single cutoffparameter. Such a network can be thought of an ensemble of unweighted graphs, each consisting of edges withweights greater than the cutoff value. We look at this network ensemble as a complex system with a temperature parameter, and refer to it as a Latent Network. Ourexperiments on a number of datasets from two different domains show that certain properties of latent networks like clustering coefficient, average shortest path,and connected components exhibit patterns that are significantly divergent from randomized networks. We explain that these patterns reflect the network phase transition as well as the existence of a community structure in document collections. Using numerical analysis,we show that we can use the aforementioned networkproperties to predicts the clustering Normalized MutualInformation (NMI) with high correlation (rho > 0.9). Finally we show that our clustering method significantlyoutperforms other baseline methods (NMI > 0.5)

【Keywords】:

146. Integrating Clustering and Multi-Document Summarization by Bi-Mixture Probabilistic Latent Semantic Analysis (PLSA) with Sentence Bases.

Paper Link】 【Pages】:

【Authors】: Chao Shen ; Tao Li ; Chris H. Q. Ding

【Abstract】: Probabilistic Latent Semantic Analysis (PLSA) has been popularly used in document analysis. However, as it is currently formulated, PLSA strictly requires the number of word latent classes to be equal to the number of document latent classes. In this paper, we propose Bi-mixture PLSA, a new formulation of PLSA that allows the number of latent word classes to be different from the number of latent document classes. We further extend Bi-mixture PLSA to incorporate the sentence information, and propose Bi-mixture PLSA with sentence bases (Bi-PLSAS) to simultaneously cluster and summarize the documents utilizing the mutual influence of the document clustering and summarization procedures. Experiments on real-world datasets demonstrate the effectiveness of our proposed methods.

【Keywords】:

147. Tree Sequence Kernel for Natural Language.

Paper Link】 【Pages】:

【Authors】: Jun Sun ; Min Zhang ; Chew Lim Tan

【Abstract】: We propose Tree Sequence Kernel (TSK), which implicitly exhausts the structure features of a sequence of subtrees embedded in the phrasal parse tree. By incorporating the capability of sequence kernel, TSK enriches tree kernel with tree sequence features so that it may provide additional useful patterns for machine learning applications. Two approaches of penalizing the substructures are proposed and both can be accomplished by efficient algorithms via dynamic programming. Evaluations are performed on two natural language tasks, i.e. Question Classification and Relation Extraction. Experimental results suggest that TSK outperforms tree kernel for both tasks, which also reveals that the structure features made up of multiple subtrees are effective and play a complementary role to the single tree structure.

【Keywords】:

148. WikiSimple: Automatic Simplification of Wikipedia Articles.

Paper Link】 【Pages】:

【Authors】: Kristian Woodsend ; Mirella Lapata

【Abstract】: Text simplification aims to rewrite text into simpler versions and thus make information accessible to a broader audience (e.g., non-native speakers, children, and individuals with language impairments). In this paper, we propose a model that simplifies documents automatically while selecting their most important content and rewriting them in a simpler style. We learn content selection rules from same-topic Wikipedia articles written in the main encyclopedia and its Simple English variant. We also use the revision histories of Simple Wikipedia articles to learn a quasi-synchronous grammar of simplification rewrite rules. Based on an integer linear programming formulation, we develop a joint model where preferences based on content and style are optimized simultaneously. Experiments on simplifying main Wikipedia articles show that our method significantly reduces the reading difficulty, while still capturing the important content.

【Keywords】:

149. Identifying Evaluative Sentences in Online Discussions.

Paper Link】 【Pages】:

【Authors】: Zhongwu Zhai ; Bing Liu ; Lei Zhang ; Hua Xu ; Peifa Jia

【Abstract】: Much of opinion mining research focuses on product reviews because reviews are opinion-rich and contain little irrelevant information. However, this cannot be said about online discussions and comments. In such postings, the discussions can get highly emotional and heated with many emotional statements, and even personal attacks. As a result, many of the postings and sentences do not express positive or negative opinions about the topic being discussed. To find people’s opinions on a topic and its different aspects, which we call evaluative opinions, those irrelevant sentences should be removed. The goal of this research is thus to identify evaluative opinion sentences. A novel unsupervised approach is proposed to solve the problem, and our experimental results show that it performs well.

【Keywords】:

Reasoning about Plans, Processes, and Actions 14

150. Planning for Operational Control Systems with Predictable Exogenous Events.

Paper Link】 【Pages】:

【Authors】: Ronen I. Brafman ; Carmel Domshlak ; Yagil Engel ; Zohar Feldman

【Abstract】: Various operational control systems (OCS) are naturally modeled as Markov Decision Processes. OCS often enjoy access to predictions of future events that have substantial impact on their operations. For example, reliable forecasts of extreme weather conditions are widely available, and such events can affect typical request patterns for customer response management systems, the flight and service time of airplanes, or the supply and demand patterns for electricity. The space of exogenous events impacting OCS can be very large, prohibiting their modeling within the MDP; moreover, for many of these exogenous events there is no useful predictive, probabilistic model. Realtime predictions, however, possibly with a short lead-time, are often available. In this work we motivate a model which combines offline MDP infinite horizon planning with realtime adjustments given specific predictions of future exogenous events, and suggest a framework in which such predictions are captured and trigger real-time planning problems. We propose a number of variants of existing MDP solution algorithms, adapted to this context, and evaluate them empirically.

【Keywords】:

151. Generating Diverse Plans Using Quantitative and Qualitative Plan Distance Metrics.

Paper Link】 【Pages】:

【Authors】: Alexandra Coman ; Hector Muñoz-Avila

【Abstract】: Diversity-aware planning consists of generating multiple plans which, while solving the same problem, are dissimilar from one another. Quantitative plan diversity is domain-independent and does not require extensive knowledge-engineering effort, but can fail to reflect plan differences that are relevant to users. Qualitative plan diversity is based on domain-specific characteristics, thus being of greater practical value, but may require substantial knowledge engineering. We demonstrate a domain-independent diverse plan generation method that is based on customizable plan distance metrics and amenable to both quantitative and qualitative diversity. Qualitative plan diversity is obtained with minimal knowledge-engineering effort, using distance metrics which incorporate domain-specific content.

【Keywords】:

152. A POMDP Model of Eye-Hand Coordination.

Paper Link】 【Pages】:

【Authors】: Tom Erez ; Julian J. Tramper ; William D. Smart ; Stan C. A. M. Gielen

【Abstract】: This paper presents a generative model of eye-hand coordination. We use numerical optimization to solve for the joint behavior of an eye and two hands, deriving a predicted motion pattern from first principles, without imposing heuristics. We model the planar scene as a POMDP with 17 continuous state dimensions. Belief-space optimization is facilitated by using a nominal-belief heuristic, whereby we assume (during planning) that the maximum likelihood observation is always obtained. Since a globally-optimal solution for such a high-dimensional domain is computationally intractable, we employ local optimization in the belief domain. By solving for a locally-optimal plan through belief space, we generate a motion pattern of mutual coordination between hands and eye: the eye's saccades disambiguate the scene in a task-relevant manner, and the hands' motions anticipate the eye's saccades. Finally, the model is validated through a behavioral experiment, in which human subjects perform the same eye-hand coordination task. We show how simulation is congruent with the experimental results.

【Keywords】:

153. Recognizing Plans with Loops Represented in a Lexicalized Grammar.

Paper Link】 【Pages】:

【Authors】: Christopher W. Geib ; Robert P. Goldman

【Abstract】: This paper extends existing plan recognition research to handle plans containing loops. We supply an encoding of plans with loops for recognition, based on techniques used to parse lexicalized grammars, and demonstrate its effectiveness empirically. To do this, the paper first shows how encoding plan libraries as context free grammars permits the application of standard rewriting techniques to remove left recursion and ε-productions, thereby enabling polynomial time parsing. However, these techniques alone fail to provide efficient algorithms for plan recognition. We show how the loop-handling methods from formal grammars can be extended to the more general plan recognition problem and provide a method for encoding loops in an existing plan recognition system that scales linearly in the number of loop iterations.

【Keywords】:

154. A Switching Planner for Combined Task and Observation Planning.

Paper Link】 【Pages】:

【Authors】: Moritz Göbelbecker ; Charles Gretton ; Richard Dearden

【Abstract】: From an automated planning perspective the problem of practical mobile robot control in realistic environments poses many important and contrary challenges. On the one hand, the planning process must be lightweight, robust, and timely. Over the lifetime of the robot it must always respond quickly with new plans that accommodate exogenous events, changing objectives, and the underlying unpredictability of the environment. On the other hand, in order to promote efficient behaviours the planning process must perform computationally expensive reasoning about contingencies and possible revisions of subjective beliefs according to quantitatively modelled uncertainty in acting and sensing. Towards addressing these challenges, we develop a continual planning approach that switches between using a fast satisficing "classical" planner, to decide on the overall strategy, and decision-theoretic planning to solve small abstract subproblems where deeper consideration of the sensing model is both practical, and can significantly impact overall performance. We evaluate our approach in large problems from a realistic robot exploration domain.

【Keywords】:

155. Exploiting Path Refinement Abstraction in Domain Transition Graphs.

Paper Link】 【Pages】:

【Authors】: Peter Gregory ; Derek Long ; Craig McNulty ; Susan M. Murphy

【Abstract】: Partial Refinement A-Star (PRA is an abstraction technique, based on clustering nearby nodes in graphs, useful in large path-planning problems. Abstracting the underlying graph yields a simpler problem whose solution can be used, by refinement, as a guide to a solution to the original problem. A fruitful way to view domain independent planning problems is as a collection of multi-valued variables that must perform synchronised transitions through graphs of possible values, where the edges are defined by the domain actions. Planning involves finding efficient paths through Domain Transition Graphs (DTGs). In problems where these graphs are large, planning can be prohibitively expensive. In this paper we explore two ways to exploit PRA in DTGs.

【Keywords】:

156. The Inter-League Extension of the Traveling Tournament Problem and its Application to Sports Scheduling.

Paper Link】 【Pages】:

【Authors】: Richard Hoshino ; Ken-ichi Kawarabayashi

【Abstract】: With the recent inclusion of inter-league games to professional sports leagues, a natural question is to determine the "best possible" inter-league schedule that retains all of the league's scheduling constraints to ensure competitive balance and fairness, while minimizing the total travel distance for both economic and environmental efficiency. To answer that question, this paper introduces the Bipartite Traveling Tournament Problem (BTTP) , the inter-league extension of the well-studied Traveling Tournament Problem. We prove that the 2n -team BTTP is NP-complete, but for small values of n , a distance-optimal inter-league schedule can be generated from an algorithm based on minimum-weight 4-cycle-covers. We apply our algorithm to the 12-team Nippon Professional Baseball (NPB) league in Japan, creating an inter-league tournament that reduces total team travel by 16% compared to the actual schedule played by these teams during the 2010 NPB season. We also analyze the problem of inter-league scheduling for the 30-team National Basketball Association (NBA), and develop a tournament schedule whose total inter-league travel distance is just 3.8% higher than the trivial theoretical lower bound.  

【Keywords】:

Paper Link】 【Pages】:

【Authors】: Tatsuya Imai ; Akihiro Kishimoto

【Abstract】: Greedy best-first search (GBFS) is a popular and effective algorithm in satisficing planning and is incorporated into high-performance planners. GBFS in planning decides its search direction with automatically generated heuristic functions. However, if the heuristic functions evaluate nodes inaccurately, GBFS may be misled into a valueless search direction, thus resulting in performance degradation. This paper presents a simple but effective algorithm considering a diversity of search directions to avoid the errors of heuristic information. Experimental results in solving a variety of planning problems show that our approach is successful.

【Keywords】:

158. Improving Cost-Optimal Domain-Independent Symbolic Planning.

Paper Link】 【Pages】:

【Authors】: Peter Kissmann ; Stefan Edelkamp

【Abstract】: Symbolic search with BDDs has shown remarkable performance for cost-optimal deterministic planning by exploiting a succinct representation and exploration of state sets. In this paper we enhance BDD-based planning by applying a combination of domain-independent search techniques: the optimization of the variable ordering in the BDD by approximating the linear arrangement problem, pattern selection for improved construction of search heuristics in form of symbolic partial pattern databases, and a decision procedure for the amount of bidirection in the symbolic search process.

【Keywords】:

159. On Improving Conformant Planners by Analyzing Domain-Structures.

Paper Link】 【Pages】:

【Authors】: Hoang-Khoi Nguyen ; Dang-Vien Tran ; Tran Cao Son ; Enrico Pontelli

【Abstract】: The paper introduces a novel technique for improving the performance and scalability of best-first progression-based conformant planners. The technique is inspired by different well-known techniques from classical planning, such as landmark and stratification. Its most salient feature is that it is relatively cheap to implement yet quite effective when applicable. The effectiveness of the proposed technique is demonstrated by the development of new conformant planners by integrating the technique in various state-of-the-art conformant planners and an extensive experimental evaluation of the new planners using benchmarks collected from various sources. The result shows that the technique can be applied in several benchmarks and helps improve both performance and scalability of conformant planners.

【Keywords】:

160. Exploiting Problem Symmetries in State-Based Planners.

Paper Link】 【Pages】:

【Authors】: Nir Pochter ; Aviv Zohar ; Jeffrey S. Rosenschein

【Abstract】: Previous research in Artificial Intelligence has identified the possibility of simplifying planning problems via the identification and exploitation of symmetries. We advance the state of the art in algorithms that exploit symmetry in planning problems by generalizing previous approaches, and applying symmetry reductions to state-based planners. We suggest several algorithms for symmetry exploitation in state-based search, but also provide a comprehensive view through which additional algorithms can be developed and fine-tuned. We evaluate our approach to symmetry exploitation on instances from previous planning competitions, and demonstrate that our algorithms significantly improve the solution time of instances with symmetries.

【Keywords】:

161. Qualitative Numeric Planning.

Paper Link】 【Pages】:

【Authors】: Siddharth Srivastava ; Shlomo Zilberstein ; Neil Immerman ; Hector Geffner

【Abstract】: We consider a new class of planning problems involving a set of non-negative real variables, and a set of non-deterministic actions that increase or decrease the values of these variables by some arbitrary amount. The formulas specifying the initial state, goal state, or action preconditions can only assert whether certain variables are equal to zero or not. Assuming that the state of the variables is fully observable, we obtain two results. First, the solution to the problem can be expressed as a policy mapping qualitative states into actions, where a qualitative state includes a Boolean variable for each original variable, indicating whether its value is zero or not. Second, testing whether any such policy, that may express nested loops of actions, is a solution to the problem, can be determined in time that is polynomial in the qualitative state space, which is much smaller than the original infinite state space. We also report experimental results using a simple generate-and-test planner to illustrate these findings.

【Keywords】:

162. Extending Classical Planning Heuristics to Probabilistic Planning with Dead-Ends.

Paper Link】 【Pages】:

【Authors】: Florent Teichteil-Königsbuch ; Vincent Vidal ; Guillaume Infantes

【Abstract】: Recent domain-determinization techniques have been very successful in many probabilistic planning problems. We claim that traditional heuristic MDP algorithms have been unsuccessful due mostly to the lack of efficient heuristics in structured domains. Previous attempts like mGPT used classical planning heuristics to an all-outcome determinization of MDPs without discount factor; yet, discounted optimization is required to solve problems with potential dead-ends. We propose a general extension of classical planning heuristics to goal-oriented discounted MDPs, in order to overcome this flaw. We apply our theoretical analysis to the well-known classical planning heuristics Hmax and Hadd, and prove that the extended Hmax is admissible. We plugged our extended heuristics to popular graph-based (Improved-LAO, LRTDP, LDFS) and ADD-based (sLAO, sRTDP) MDP algorithms: experimental evaluations highlight competitive results compared with the winners of previous competitions (FF-Replan, FPG, RFF), and show that our discounted heuristics solve more problems than non-discounted ones, with better criteria values. As for classical planning, the extended Hadd outperforms the extended Hmax on most problems.

【Keywords】:

163. Conjunctive Representations in Contingent Planning: Prime Implicates Versus Minimal CNF Formula.

Paper Link】 【Pages】:

【Authors】: Son Thanh To ; Tran Cao Son ; Enrico Pontelli

【Abstract】: This paper compares in depth the effectiveness of two conjunctive belief state representations in contingent planning: prime implicates and minimal CNF, a compact form of CNF formulae, which were initially proposed in conformant planning research (To et al. 2010a; 2010b). Similar to the development of the contingent planner CNFct for minimal CNF (To et al. 2011b), the present paper extends the progression function for the prime implicate representation in (To et al. 2010b) for computing successor belief states in the presence of incomplete information to handle non-deterministic and sensing actions required in contingent planning. The idea was instantiated in a new contingent planner, called PIct, using the same AND/OR search algorithm and heuristic function as those for CNFct. The experiments show that, like CNFct, PIct performs very well in a wide range of benchmarks. The study investigates the advantages and disadvantages of the two planners and identifies the properties of each representation method that affect the performance.

【Keywords】:

Reasoning under Uncertainty 9

164. Efficient Methods for Lifted Inference with Aggregate Factors.

Paper Link】 【Pages】:

【Authors】: Jaesik Choi ; Rodrigo de Salvo Braz ; Hung Hai Bui

【Abstract】: Aggregate factors (that is, those based on aggregate functions such as SUM, AVERAGE, AND etc.) in probabilistic relational models can compactly represent dependencies among a large number of relational random variables. However, propositional inference on a factor aggregating n k -valued random variables into an r -valued result random variable is O ( r k 2 n ). Lifted methods can ameliorate this to O ( r n k ) in general and O ( r k log n ) for commutative associative aggregators. In this paper, we propose (a) an exact solution constant in n when k  = 2 for certain aggregate operations such as AND, OR and SUM, and (b) a close approximation for inference with aggregate factors with time complexity constant in n . This approximate inference involves an analytical solution for some operations when k > 2. The approximation is based on the fact that the typically used aggregate functions can be represented by linear constraints in the standard ( k  –1)-simplex in R k where k is the number of possible values for random variables. This includes even aggregate functions that are commutative but not associative (e.g., the MODE operator that chooses the most frequent value). Our algorithm takes polynomial time in k (which is only 2 for binary variables) regardless of r and n, and the error decreases as n increases. Therefore, for most applications (in which a close approximation suffices) our algorithm is a much more efficient solution than existing algorithms. We present experimental results supporting these claims. We also present a (c) third contribution which further optimizes aggregations over multiple groups of random variables with distinct distributions.

【Keywords】:

165. Dual Decomposition for Marginal Inference.

Paper Link】 【Pages】:

【Authors】: Justin Domke

【Abstract】: We present a dual decomposition approach to the tree-reweighted belief propagation objective. Each tree in the tree-reweighted bound yields one subproblem, which can be solved with the sum-product algorithm. The master problem is a simple differentiable optimization, to which a standard optimization method can be applied. Experimental results on 10x10 Ising models show the dual decomposition approach using L-BFGS is similar in settings where message-passing converges quickly, and one to two orders of magnitude faster in settings where message-passing requires many iterations, specifically high accuracy convergence, and strong interactions.

【Keywords】:

166. Stopping Rules for Randomized Greedy Triangulation Schemes.

Paper Link】 【Pages】:

【Authors】: Andrew Gelfand ; Kalev Kask ; Rina Dechter

【Abstract】: Many algorithms for performing inference in graphical models have complexity that is exponential in the treewidth — a parameter of the underlying graph structure. Computing the (minimal) treewidth is NPcomplete, so stochastic algorithms are sometimes used to find low width tree decompositions. A common approach for finding good decompositions is iteratively executing a greedy triangulation algorithm (e.g. minfill) with randomized tie-breaking. However, utilizing a stochastic algorithm as part of the inference task introduces a new problem — namely, deciding how long the stochastic algorithm should be allowed to execute before performing inference on the best tree decomposition found so far. We refer to this dilemma as the Stopping Problem and formalize it in terms of the total time needed to answer a probabilistic query. We propose a rule for discontinuing the search for improved decompositions and demonstrate the benefit (in terms of time saved) of applying this rule to Bayes and Markov network instances.

【Keywords】:

167. Coarse-to-Fine Inference and Learning for First-Order Probabilistic Models.

Paper Link】 【Pages】:

【Authors】: Chloé Kiddon ; Pedro M. Domingos

【Abstract】: Coarse-to-fine approaches use sequences of increasingly fine approximations to control the complexity of inference and learning. These techniques are often used in NLP and vision applications. However, no coarse-to-fine inference or learning methods have been developed for general first-order probabilistic domains, where the potential gains are even higher. We present our Coarse-to-Fine Probabilistic Inference (CFPI) framework for general coarse-to-fine inference for first-order probabilistic models, which leverages a given or induced type hierarchy over objects in the domain. Starting by considering the inference problem at the coarsest type level, our approach performs inference at successively finer grains, pruning high- and low-probability atoms before refining. CFPI can be applied with any probabilistic inference method and can be used in both propositional and relational domains. CFPI provides theoretical guarantees on the errors incurred, and these guarantees can be tightened when CFPI is applied to specific inference algorithms. We also show how to learn parameters in a coarse-to-fine manner to maximize the efficiency of CFPI. We evaluate CFPI with the lifted belief propagation algorithm on social network link prediction and biomolecular event prediction tasks. These experiments show CFPI can greatly speed up inference without sacrificing accuracy.

【Keywords】:

168. Memory-Efficient Dynamic Programming for Learning Optimal Bayesian Networks.

Paper Link】 【Pages】:

【Authors】: Brandon M. Malone ; Changhe Yuan ; Eric A. Hansen

【Abstract】: We describe a memory-efficient implementation of a dynamic programming algorithm for learning the optimal structure of a Bayesian network from training data. The algorithm leverages the layered structure of the dynamic programming graphs representing the recursive decomposition of the problem to reduce the memory requirements of the algorithm from O(n2 n ) to O(C(n, n/2)), where C(n, n/2) is the binomial coefficient. Experimental results show that the approach runs up to an order of magnitude faster and scales to datasets with more variables than previous approaches.

【Keywords】:

169. When to Stop? That Is the Question.

Paper Link】 【Pages】:

【Authors】: Shulamit Reches ; Meir Kalech ; Rami Stern

【Abstract】: When to make a decision is a key question in decision making problems characterized by uncertainty. In this paper we deal with decision making in environments where the information arrives dynamically. We address the tradeoff between waiting and stopping strategies. On the one hand, waiting to obtain more information reduces the uncertainty, but it comes with a cost. On the other hand, stopping and making a decision based on an expected utility, decreases the cost of waiting, but the decision is made based on uncertain information. In this paper, we prove that computing the optimal time to make a decision that guarantees the optimal utility is NP-hard. We propose a pessimistic approximation that guarantees an optimal decision when the recommendation is to wait. We empirically evaluate our algorithm and show that the quality of the decision is near-optimal and much faster than the optimal algorithm.

【Keywords】:

170. Abductive Markov Logic for Plan Recognition.

Paper Link】 【Pages】:

【Authors】: Parag Singla ; Raymond J. Mooney

【Abstract】: Plan recognition is a form of abductive reasoning that involves inferring plans that best explain sets of observed actions. Most existing approaches to plan recognition and other abductive tasks employ either purely logical methods that donot handle uncertainty, or purely probabilistic methods thatdo not handle structured representations. To overcome these limitations, this paper introduces an approach to abductive reasoning using a first-order probabilistic logic, specifically Markov Logic Networks (MLNs). It introduces several novel techniques for making MLNs efficient and effective for abduction. Experiments on three plan recognition datasets showthe benefit of our approach over existing methods.

【Keywords】:

171. Fast Parallel and Adaptive Updates for Dual-Decomposition Solvers.

Paper Link】 【Pages】:

【Authors】: Özgür Sümer ; Umut A. Acar ; Alexander T. Ihler ; Ramgopal R. Mettu

【Abstract】: Dual-decomposition (DD) methods are quickly becoming important tools for estimating the minimum energy state of a graphical model. DD methods decompose a complex model into a collection of simpler subproblems that can be solved exactly (such as trees), that in combination provide upper and lower bounds on the exact solution. Subproblem choice can play a major role: larger subproblems tend to improve the bound more per iteration, while smaller subproblems enable highly parallel solvers and can benefit from re-using past solutions when there are few changes between iterations. We propose an algorithm that can balance many of these aspects to speed up convergence. Our method uses a cluster tree data structure that has been proposed for adaptive exact inference tasks, and we apply it in this paper to dual-decomposition approximate inference. This approach allows us to process large subproblems to improve the bounds at each iteration, while allowing a high degree of parallelizability and taking advantage of subproblems with sparse updates. For both synthetic inputs and a real-world stereo matching problem, we demonstrate that our algorithm is able to achieve significant improvement in convergence time.

【Keywords】:

172. Utilizing Partial Policies for Identifying Equivalence of Behavioral Models.

Paper Link】 【Pages】:

【Authors】: Yifeng Zeng ; Prashant Doshi ; Yinghui Pan ; Hua Mao ; Muthukumaran Chandrasekaran ; Jian Luo

【Abstract】: We present a novel approach for identifying exact and approximate behavioral equivalence between models of agents. This is significant because both decision making and game play in multiagent settings must contend with behavioral models of other agents in order to predict their actions. One approach that reduces the complexity of the model space is to group models that are behaviorally equivalent. Identifying equivalence between models requires solving them and comparing entire policy trees. Because the trees grow exponentially with the horizon, our approach is to focus on partial policy trees for comparison and determining the distance between updated beliefs at the leaves of the trees. We propose a principled way to determine how much of the policy trees to consider, which trades off solution quality for efficiency. We investigate this approach in the context of the interactive dynamic influence diagram and evaluate its performance.

【Keywords】:

Robotics 7

173. Multiagent Patrol Generalized to Complex Environmental Conditions.

Paper Link】 【Pages】:

【Authors】: Noa Agmon ; Daniel Urieli ; Peter Stone

【Abstract】: The problem of multiagent patrol has gained considerable attention during the past decade, with the immediate applicability of the problem being one of its main sources of interest. In this paper we concentrate on frequency-based patrol, in which the agents' goal is to optimize a frequency criterion, namely, minimizing the time between visits to a set of interest points. We consider multiagent patrol in environments with complex environmental conditions that affect the cost of traveling from one point to another. For example, in marine environments, the travel time of ships depends on parameters such as wind, water currents, and waves. We demonstrate that in such environments there is a need to consider a new multiagent patrol strategy which divides the given area into parts in which more than one agent is active, for improving frequency. We show that in general graphs this problem is intractable, therefore we focus on simplified (yet realistic) cyclic graphs with possible inner edges. Although the problem remains generally intractable in such graphs, we provide a heuristic algorithm that is shown to significantly improve point-visit frequency compared to other patrol strategies. For evaluation of our work we used a custom developed ship simulator that realistically models ship movement constraints such as engine force and drag and reaction of the ship to environmental changes.

【Keywords】:

174. Automated Abstractions for Patrolling Security Games.

Paper Link】 【Pages】:

【Authors】: Nicola Basilico ; Nicola Gatti

【Abstract】: Recently, there has been a significant interest in studying security games to provide tools for addressing resource allocation problems in security applications. Patrolling security games (PSGs) constitute a special class of security games wherein the resources are mobile. One of the most relevant open problems in security games is the design of scalable algorithms to tackle realistic scenarios. While the literature mainly focuses on heuristics and decomposition techniques (e.g., double oracle), in this paper we provide, to the best of our knowledge, the first study on the use of abstractions in security games (specifically for PSGs) to design scalable algorithms. We define some classes of abstractions and we provide parametric algorithms to automatically generate abstractions. We show that abstractions allow one to relax the constraint of patrolling strategies' Markovianity (customary in PSGs) and to solve large game instances. We additionally pose the problem to search for the optimal abstraction and we develop an anytime algorithm to find it.

【Keywords】:

175. Comparing Action-Query Strategies in Semi-Autonomous Agents.

Paper Link】 【Pages】:

【Authors】: Robert Cohn ; Edmund H. Durfee ; Satinder P. Singh

【Abstract】: We consider settings in which a semi-autonomous agent has uncertain knowledge about its environment, but can ask what action the human operator would prefer taking in the current or in a potential future state. Asking queries can improve behavior, but if queries come at a cost (e.g., due to limited operator attention), the value of each query should be maximized. We compare two strategies for selecting action queries: 1) based on myopically maximizing expected gain in long-term value, and 2) based on myopically minimizing uncertainty in the agent's policy representation. We show empirically that the first strategy tends to select more valuable queries, and that a hybrid method can outperform either method alone in settings with limited computation.

【Keywords】:

176. Optimal Route Planning for Electric Vehicles in Large Networks.

Paper Link】 【Pages】:

【Authors】: Jochen Eisner ; Stefan Funke ; Sabine Storandt

【Abstract】: We consider the problem of routing electric vehicles (EV) in the most energy-efficient way within a road network taking into account both their limited energy supply as well as their ability to recuperate energy. Employing a classical result by Johnson and an observation about Dijkstra under non-constant edge costs we obtain O(n log n +m) query time after a O(nm) preprocessing phase for any road network graph whose edge costs represent energy consumption or recuperation.If the energy recuperation is height induced in a very natural way,the preprocessing phase can even be omitted. We then adapt a technique for speeding-up (unconstrained) shortest path queries to our scenario to achieve a speed-up of another factor of around 20. Our results drastically improve upon the recent results in (Artmeier et al. 2010) and allow for route planning of EVs in an instant even on large networks.

【Keywords】:

177. Online Graph Pruning for Pathfinding On Grid Maps.

Paper Link】 【Pages】:

【Authors】: Daniel Damir Harabor ; Alban Grastien

【Abstract】: Pathfinding in uniform-cost grid environments is a problem commonly found in application areas such as robotics and video games. The state-of-the-art is dominated by hierarchical pathfinding algorithms which are fast and have small memory overheads but usually return suboptimal paths. In this paper we present a novel search strategy, specific to grids, which is fast, optimal and requires no memory overhead. Our algorithm can be described as a macro operator which identifies and selectively expands only certain nodes in a grid map which we call jump points. Intermediate nodes on a path connecting two jump points are never expanded. We prove that this approach always computes optimal solutions and then undertake a thorough empirical analysis, comparing our method with related works from the literature. We find that searching with jump points can speed up A* by an order of magnitude and more and report significant improvement over the current state of the art.

【Keywords】:

178. Complete Information Pursuit Evasion in Polygonal Environments.

Paper Link】 【Pages】:

【Authors】: Kyle Klein ; Subhash Suri

【Abstract】: Suppose an unpredictable evader is free to move around in a polygonal environment of arbitrary complexity that is under full camera surveillance. How many pursuers, each with the same maximum speed as the evader, are necessary and sufficient to guarantee a successful capture of the evader? The pursuers always know the evader's current position through the camera network, but need to physically reach the evader to capture it. We allow the evader the knowledge of the current positions of all the pursuers as well---this accords with the standard worst-case analysis model, but also models a practical situation where the evader has ``hacked'' into the surveillance system. Our main result is to prove that three pursuers are always sufficient and sometimes necessary to capture the evader. The bound is independent of the number of vertices or holes in the polygonal environment.

【Keywords】:

179. Learning Dimensional Descent for Optimal Motion Planning in High-dimensional Spaces.

Paper Link】 【Pages】:

【Authors】: Paul Vernaza ; Daniel D. Lee

【Abstract】: We present a novel learning-based method for generating optimal motion plans for high-dimensional motion planning problems. In order to cope with the curse of dimensional- ity, our method proceeds in a fashion similar to block co- ordinate descent in finite-dimensional optimization: at each iteration, the motion is optimized over a lower dimensional subspace while leaving the path fixed along the other dimen- sions. Naive implementations of such an idea can produce vastly suboptimal results. In this work, we show how a prof- itable set of directions in which to perform this dimensional descent procedure can be learned efficiently. We provide suf- ficient conditions for global optimality of dimensional de- scent in this learned basis, based upon the low-dimensional structure of the planning cost function. We also show how this dimensional descent procedure can easily be used for problems that do not exhibit such structure with monotonic convergence. We illustrate the application of our method to high dimensional shape planning and arm trajectory planning problems.

【Keywords】:

Special Track on AI and the Web 28

180. Detecting Multilingual and Multi-Regional Query Intent in Web Search.

Paper Link】 【Pages】:

【Authors】: Yi Chang ; Ruiqiang Zhang ; Srihari Reddy ; Yan Liu

【Abstract】: With rapid growth of commercial search engines, detecting multilingual and multi-regional intent underlying search queries becomes a critical challenge to serve international users with diverse language and region requirements. We introduce a query intent probabilistic model, whose input is the number of clicks on documents from different regions and in different language, while the output of this model is a smoothed probabilistic distribution of multilingual and multi-regional query intent. Based on an editorial test to evaluate the accuracy of the intent classifier, our probabilistic model could improve the accuracy of multilingual intent detection for 15%, and improve multi-regional intent detection for 18%. To improve web search quality, we propose a set of new ranking features to combine multilingual and multi-regional query intent with document language/region attributes, and apply different approaches in integrating intent information to directly affect ranking. The experiments show that the novel features could provide 2.31% NDCG@1 improvement and 1.81% NDCG@5 improvement.

【Keywords】:

Paper Link】 【Pages】:

【Authors】: Weizhu Chen ; Zhanglong Ji ; Si Shen ; Qiang Yang

【Abstract】: Recent advances in click modeling have established it as an attractive approach to interpret search click data. These advances characterize users’ search behavior either in advertisement blocks, or within an organic search block through probabilistic models. Yet, when searching for information on a search result page, one is often interacting with the search engine via an entire page instead of a single block. Consequently, previous works that exclusively modeled user behavior in a single block may sacrifice much useful user behavior information embedded in other blocks. To solve this problem, in this paper, we put forward a novel Whole Page Click (WPC) Model to characterize user behavior in multiple blocks. Specifically, WPC uses a Markov chain to learn the user transition probabilities among different blocks in the whole page. To compare our model with the best alternatives in the Web-Search literature, we run a large-scale experiment on a real dataset and demonstrate the advantage of the WPC model in terms of both the whole page and each block in the page. Especially, we find that WPC can achieve significant gain in interpreting the advertisement data, despite of the sparsity of the advertisement click data.

【Keywords】:

182. User-Controllable Learning of Location Privacy Policies With Gaussian Mixture Models.

Paper Link】 【Pages】:

【Authors】: Justin Cranshaw ; Jonathan Mugan ; Norman M. Sadeh

【Abstract】: With smart-phones becoming increasingly commonplace, there has been a subsequent surge in applications that continuously track the location of users. However, serious privacy concerns arise as people start to widely adopt these applications. Users will need to maintain policies to determine under which circumstances to share their location. Specifying these policies however, is a cumbersome task, suggesting that machine learning might be helpful. In this paper, we present a user-controllable method for learning location sharing policies. We use a classifier based on multivariate Gaussian mixtures that is suitably modified so as to restrict the evolution of the underlying policy to favor incremental and therefore human-understandable changes as new data arrives. We evaluate the model on real location-sharing policies collected from a live location-sharing social network, and we show that our method can learn policies in a user-controllable setting that are just as accurate as policies that do not evolve incrementally. Additionally, we highlight the strength of the generative modeling approach we take, by showing how our model easily extends to the semi-supervised setting.

【Keywords】:

183. Artificial Intelligence for Artificial Artificial Intelligence.

Paper Link】 【Pages】:

【Authors】: Peng Dai ; Mausam ; Daniel S. Weld

【Abstract】: Crowdsourcing platforms such as Amazon Mechanical Turk have become popular for a wide variety of human intelligence tasks; however, quality control continues to be a significant challenge. Recently, we propose TurKontrol, a theoretical model based on POMDPs to optimize iterative, crowd-sourced workflows. However, they neither describe how to learn the model parameters, nor show its effectiveness in a real crowd-sourced setting. Learning is challenging due to the scale of the model and noisy data: there are hundreds of thousands of workers with high-variance abilities. This paper presents an end-to-end system that first learns TurKontrol's POMDP parameters from real Mechanical Turk data, and then applies the model to dynamically optimize live tasks. We validate the model and use it to control a successive-improvement process on Mechanical Turk. By modeling worker accuracy and voting patterns, our system produces significantly superior artifacts compared to those generated through nonadaptive workflows using the same amount of money.

【Keywords】:

184. Towards Practical ABox Abduction in Large OWL DL Ontologies.

Paper Link】 【Pages】:

【Authors】: Jianfeng Du ; Guilin Qi ; Yi-Dong Shen ; Jeff Z. Pan

【Abstract】: ABox abduction is an important aspect for abductive reasoning in Description Logics (DLs). It finds all minimal sets of ABox axioms that should be added to a background ontology to enforce entailment of a specified set of ABox axioms. As far as we know, by now there is only one ABox abduction method in expressive DLs computing abductive solutions with certain minimality. However, the method targets an ABox abduction problem that may have infinitely many abductive solutions and may not output an abductive solution in finite time. Hence, in this paper we propose a new ABox abduction problem which has only finitely many abductive solutions and also propose a novel method to solve it. The method reduces the original problem to an abduction problem in logic programming and solves it with Prolog engines. Experimental results show that the method is able to compute abductive solutions in benchmark OWL DL ontologies with large ABoxes.

【Keywords】:

185. Identifying Missing Node Information in Social Networks.

Paper Link】 【Pages】:

【Authors】: Ron Eyal ; Sarit Kraus ; Avi Rosenfeld

【Abstract】: In recent years, social networks have surged in popularity as one of the main applications of the Internet. This has generated great interest in researching these networks by various fields in the scientific community. One key aspect of social network research is identifying important missing information which is not explicitly represented in the network, or is not visible to all. To date, this line of research typically focused on what connections were missing between nodes,or what is termed the "Missing Link Problem." This paper introduces a new Missing Nodes Identification problem where missing members in the social network structure must be identified. Towards solving this problem, we present an approach based on clustering algorithms combined with measures from missing link research. We show that this approach has beneficial results in the missing nodes identification process and we measure its performance in several different scenarios.

【Keywords】:

186. Maximum Entropy Context Models for Ranking Biographical Answers to Open-Domain Definition Questions.

Paper Link】 【Pages】:

【Authors】: Alejandro Figueroa ; John Atkinson

【Abstract】: In the context of question-answering systems, there are several strategies for scoring candidate answers to definition queries including centroid vectors, bi-term and context language models. These techniques use only positive examples (i.e., descriptions) when building their models. In this work, a maximum entropy based extension is proposed for context language models so as to account for regularities across non-descriptions mined from web-snippets. Experiments show that this extension outperforms other strategies increasing the precision of the top five ranked answers by more than 5%. Results suggest that web-snippets are a cost-efficient source of non-descriptions, and that some relationships extracted from dependency trees are effective to mine for candidate answer sentences.

【Keywords】:

187. Commonsense Causal Reasoning Using Millions of Personal Stories.

Paper Link】 【Pages】:

【Authors】: Andrew S. Gordon ; Cosmin Adrian Bejan ; Kenji Sagae

【Abstract】: The personal stories that people write in their Internet weblogs include a substantial amount of information about the causal relationships between everyday events. In this paper we describe our efforts to use millions of these stories for automated commonsense causal reasoning. Casting the commonsense causal reasoning problem as a Choice of Plausible Alternatives, we describe four experiments that compare various statistical and information retrieval approaches to exploit causal information in story corpora. The top performing system in these experiments uses a simple co-occurrence statistic between words in the causal antecedent and consequent, calculated as the Pointwise Mutual Information between words in a corpus of millions of personal stories.

【Keywords】:

188. Active Dual Collaborative Filtering with Both Item and Attribute Feedback.

Paper Link】 【Pages】:

【Authors】: Luheng He ; Nathan Nan Liu ; Qiang Yang

【Abstract】: The new user problem (aka user cold start) is very common in online recommender systems. Active collaborative filtering (active CF) tries to solve this problem by intelligently soliciting user feedback in order to build an initial user profile with minimal costs. Existing methods only query the user for feedback on items, while users can have preferences over items as well as certain item attributes. In this paper, we extend active CF via user feedback on both items and attributes. For example, when making movie recommendations, the system can ask users for not only their favorite movies, but also attributes such as genres, actors, etc. We design a unified active CF framework for incorporating both item and attribute feedback based on the random walk model. We test the active CF algorithm on real-world movie recommendation data sets to demonstrate that appropriately querying for both item and feature feedback can significantly reduce the overall user effort measured in terms of number of queries. We show that we can achieve much better recommendation quality as compared to traditional active CF methods that support only item feedback.

【Keywords】:

189. Fast Query Recommendation by Search.

Paper Link】 【Pages】:

【Authors】: Qixia Jiang ; Maosong Sun

【Abstract】: Query recommendation can not only effectively facilitate users to obtain their desired information but alsoincrease ads’ click-through rates. This paper presentsa general and highly efficient method for query recommendation. Given query sessions, we automatically generate many similar and dissimilar query-pairs as the prior knowledge. Then we learn a transformation from the prior knowledge to move similar queries closer such that similar queries tend to have similar hash values.This is formulated as minimizing the empirical error on the prior knowledge while maximizing the gap between the data and some partition hyperplanes randomly generated in advance. In the recommendation stage, we search queries that have similar hash values to the given query, rank the found queries and return the top K queries as the recommendation result. All the experimental results demonstrate that our method achieves encouraging results in terms of efficiency and recommendation performance.

【Keywords】:

190. Continual Planning with Sensing for Web Service Composition.

Paper Link】 【Pages】:

【Authors】: Eirini Kaldeli ; Alexander Lazovik ; Marco Aiello

【Abstract】: Web Service (WS) domains constitute an application field where automated planning can significantly contribute towards achieving customisable and adaptable compositions. Following the vision of using domain-independent planning and declarative complex goals to generate compositions based on atomic service descriptions, we apply a planning framework based on Constraint Satisfaction techniques to a domain consisting of WSs with diverse functionalities. One of the key requirements of such domains is the ability to address the incomplete knowledge problem, as well as recovering from failures that may occur during execution. We propose an algorithm for interleaving planning, monitoring and execution, where continual planning via altering the CSP is performed, under the light of the feedback acquired at runtime. The system is evaluated against a number of scenarios including real WSs, demonstrating the leverage of situations that can be effectively tackled with respect to previous approaches.

【Keywords】:

191. Understanding User Migration Patterns in Social Media.

Paper Link】 【Pages】:

【Authors】: Shamanth Kumar ; Reza Zafarani ; Huan Liu

【Abstract】: The incredible growth of the social web over the last decade has ushered in a flurry of new social media sites. On one hand, users have an inordinate number of choices; on the other hand, users are constrained by limited time and resources and have to choose sites in order to remain social and active. Hence, dynamic social media entails user migration, a well studied phenomenon in fields such as sociology and psychology. Users are valuable assets for social media sites as they help contribute to the growth of a site and generate revenue by increased traffic. We are intrigued to know if social media user migration can be studied, and what migration patterns are. In particular, we investigate whether people migrate, and if they do, how they migrate. We formalize site and attention migration to help identify the migration between popular social media sites and determine clear patterns of migration between sites. This work suggests a feasible way to study migration patterns in social media. The discovered patterns can help understand social media sites and gauge their popularity to improve business intelligence and revenue generation through the retention of users.

【Keywords】:

Paper Link】 【Pages】:

【Authors】: Edith Law ; Haoqi Zhang

【Abstract】: Behind every search query is a high-level mission that the user wants to accomplish.  While current search engines can often provide relevant information in response to well-specified queries, they place the heavy burden of making a plan for achieving a mission on the user. We take the alternative approach of tackling users' high-level missions directly by introducing a human computation system that generates simple plans, by decomposing a mission into goals and retrieving search results tailored to each goal. Results show that our system is able to provide users with diverse, actionable search results and useful roadmaps for accomplishing their missions.

【Keywords】:

193. Personalizing Your Web Services with Constructive DL Reasoning Join.

Paper Link】 【Pages】:

【Authors】: Freddy Lécué

【Abstract】: Nowadays web users have clearly expressed their wishes to receive and interact with personalized services directly. However, existing approaches, largely syntactic content-based, fail to provide robust, accurate and useful personalized services to its users. Towards such an issue, the semantic web provides technologies to annotate and match services’ descriptions with users’ features, interests and preferences, thus allowing for more efficient access to services and more generally information. The aim of our work, part of service personalization, is on automated instantiation of services which is crucial for advanced usability i.e., how to prepare and present services ready to be executed while limiting useless interactions with users? We introduce the constructive Description Logics reasoning join and couple it with concept abduction to i) identify useful parts of users profiles that satisfy services requirements and ii) compute the description required by a service to be executed but not provided by users profiles.

【Keywords】:

194. Trust Transitivity in Complex Social Networks.

Paper Link】 【Pages】:

【Authors】: Guanfeng Liu ; Yan Wang ; Mehmet A. Orgun

【Abstract】: In Online Social Networks (OSNs), participants can conduct rich activities, where trust is one of the most important factors for their decision making. This necessitates the evaluation of the trustworthiness between two unknown participants along the social trust paths between them based on the trust transitivity properties (i.e., if A trusts B and B trusts C, then A can trust C to some extent). In order to compute more reasonable trust value between two unknown participants, a critical and challenging problem is to make clear how and to what extent trust is transitive along a social trust path. To address this problem, we first propose a new complex social network structure that takes, besides trust, social relationships, recommendation roles and preference similarity between participants into account. These factors have significant influence on trust transitivity. We then propose a general concept, called Quality of Trust Transitivity (QoTT), that takes any factor with impact on trust transitivity as an attribute to illustrate the ability of a trust path to guarantee a certain level of quality in trust transitivity. Finally, we propose a novel Multiple QoTT Constrained Trust Transitivity (MQCTT) model. The results of our experiments demonstrate that our proposed MQCTT model follows the properties of trust and the principles illustrated in social psychology, and thus can compute more resonable trust values than existing methods that consider neither the impact of social aspects nor the properties of trust.

【Keywords】:

Paper Link】 【Pages】:

【Authors】: Hengjie Song ; Chunyan Miao ; Zhiqi Shen

【Abstract】: In current search engines, ranking functions are learned from a large number of labeled <query, URL> pairs in which the labels are assigned by human judges, describing how well the URLs match the different queries. However in commercial search engines, collecting high quality labels is time-consuming and labor-intensive. To tackle this issue, this paper studies how to produce the true relevance labels for  <query, URL> pairs using clickthrough data. By analyzing the correlations between query frequency, true relevance labels and users’ behaviors, we demonstrate that the users who search the queries with similar frequency have similar search intents and behavioral characteristics. Based on such properties, we propose an efficient discriminative parameter estimation in a multiple instance learning algorithm (MIL) to automatically produce true relevance labels for  <query, URL> pairs. Furthermore, we test our approach using a set of real world data extracted from a Chinese commercial search engine. Experimental results not only validate the effectiveness of the proposed approach, but also indicate that our approach is more likely to agree with the aggregation of the multiple judgments when strong disagreements exist in the panel of judges. In the event that the panel of judges is consensus, our approach provides more accurate automatic label results. In contrast with other models, our approach effectively improves the correlation between automatic labels and manual labels.

【Keywords】:

196. Cross-Language Latent Relational Search: Mapping Knowledge across Languages.

Paper Link】 【Pages】:

【Authors】: Nguyen Tuan Duc ; Danushka Bollegala ; Mitsuru Ishizuka

【Abstract】: Latent relational search (LRS) is a novel approach for mapping knowledge across two domains. Given a source domain knowledge concerning the Moon, "The Moon is a satellite of the Earth," one can form a question {(Moon, Earth), (Ganymede, ?)} to query an LRS engine for new knowledge in the target domain concerning the Ganymede. An LRS engine relies on some supporting sentences such as ``Ganymede is a natural satellite of Jupiter.'' to retrieve and rank "Jupiter" as the first answer. This paper proposes cross-language latent relational search (CLRS) to extend the knowledge mapping capability of LRS from cross-domain knowledge mapping to cross-domain and cross-language knowledge mapping. In CLRS, the supporting sentences for the source pair might be in a different language with that of the target pair. We represent the relation between two entities in an entity pair by lexical patterns of the context surrounding the two entities. We then propose a novel hybrid lexical pattern clustering algorithm to capture the semantic similarity between paraphrased lexical patterns across languages. Experiments on Japanese-English datasets show that the proposed method achieves an MRR of 0.579 for CLRS task, which is comparable to the MRR of an existing monolingual LRS engine.

【Keywords】:

197. Creative Introspection and Knowledge Acquisition.

Paper Link】 【Pages】:

【Authors】: Tony Veale ; Guofu Li

【Abstract】: Introspection is a question-led process in which one builds on what one already knows to explore what is possible and plausible. In creative introspection, whether in art or in science, framing the right question is as important as finding the right answer. Presupposition-laden questions are themselves a source of knowledge, and in this paper we show how widely-held beliefs about the world can be dynamically acquired by harvesting such questions from the Web. We show how metaphorical reasoning can be modeled as an introspective process, one that builds on questions harvested from the Web to pose further speculative questions and queries. Metaphor is much more than a knowledge-hungry rhetorical device: it is a conceptual lever that allows a system to extend its model of the world.

【Keywords】:

198. CCRank: Parallel Learning to Rank with Cooperative Coevolution.

Paper Link】 【Pages】:

【Authors】: Shuaiqiang Wang ; Byron J. Gao ; Ke Wang ; Hady Wirawan Lauw

【Abstract】: We propose CCRank, the first parallel algorithm for learning to rank, targeting simultaneous improvement in learning accuracy and efficiency. CCRank is based on cooperative coevolution (CC), a divide-and-conquer framework that has demonstrated high promise in function optimization for problems with large search space and complex structures. Moreover, CC naturally allows parallelization of sub-solutions to the decomposed subproblems, which can substantially boost learning efficiency. With CCRank, we investigate parallel CC in the context of learning to rank. Extensive experiments on benchmarks in comparison with the state-of-the-art algorithms show that CCRank gains in both accuracy and efficiency.

【Keywords】:

199. Integrating Community Question and Answer Archives.

Paper Link】 【Pages】:

【Authors】: Wei Wei ; Gao Cong ; Xiaoli Li ; See-Kiong Ng ; Guohui Li

【Abstract】: Question and answer pairs in Community Question Answering (CQA) services are organized into hierarchical structures or taxonomies to facilitate users to find the answers for their questions conveniently. We observe that different CQA services have their own knowledge focus and used different taxonomies to organize their question and answer pairs in their archives. As there are no simple semantic mappings between the taxonomies of the CQA services, the integration of CQA services is a challenging task. The existing approaches on integrating taxonomies ignore the hierarchical structures of the source taxonomy. In this paper, we propose a novel approach that is capable of incorporating the parent-child and sibling information in the hierarchical structures of the source taxonomy for accurate taxonomy integration. Our experimental results with real world CQA data demonstrate that the proposed method significantly outperforms state-of-the-art methods.

【Keywords】:

200. Predicting Author Blog Channels with High Value Future Posts for Monitoring.

Paper Link】 【Pages】:

【Authors】: Shanchan Wu ; Tamer Elsayed ; William Rand ; Louiqa Raschid

【Abstract】: The phenomenal growth of social media, both in scale and importance, has created a unique opportunity to track information diffusion and the spread of influence, but can also make efficient tracking difficult. Given data streams representing blog posts on multiple blog channels and a focal query post on some topic of interest, our objective is to predict which of those channels are most likely to contain a future post that is relevant, or similar, to the focal query post. We denote this task as the future author prediction problem (FAPP). This problem has applications in information diffusion for brand monitoring and blog channel personalization and recommendation. We develop prediction methods inspired by (naive) information retrieval approaches that use historical posts in the blog channel for prediction. We also train a ranking support vector machine (SVM) to solve the problem. We evaluate our methods on an extensive social media dataset; despite the difficulty of the task, all methods perform reasonably well. Results show that ranking SVM prediction can exploit blog channel and diffusion characteristics to improve prediction accuracy. Moreover, it is surprisingly good for prediction in emerging topics and identifying inconsistent authors.

【Keywords】:

201. SemRec: A Semantic Enhancement Framework for Tag Based Recommendation.

Paper Link】 【Pages】:

【Authors】: Guandong Xu ; Yanhui Gu ; Peter Dolog ; Yanchun Zhang ; Masaru Kitsuregawa

【Abstract】: Collaborative tagging services provided by various social web sites become popular means to mark web resources for different purposes such as categorization, expression of a preference and so on. However, the tags are of syntactic nature, in a free style and do not reflect semantics, resulting in the problems of redundancy, ambiguity and less semantics. Current tag-based recommender systems mainly take the explicit structural information among users, resources and tags into consideration, while neglecting the important implicit semantic relationships hidden in tagging data. In this study, we propose a Semantic Enhancement Recommendation strategy (SemRec), based on both structural information and semantic information through a unified fusion model. Extensive experiments conducted on two real datasets demonstarte the effectiveness of our approaches.

【Keywords】:

202. Analyzing and Predicting Not-Answered Questions in Community-based Question Answering Services.

Paper Link】 【Pages】:

【Authors】: Lichun Yang ; Shenghua Bao ; Qingliang Lin ; Xian Wu ; Dingyi Han ; Zhong Su ; Yong Yu

【Abstract】: This paper focuses on analyzing and predicting not-answered questions in Community based Question Answering (CQA) services, such as Yahoo! Answers. In CQA services, users express their information needs by submitting natural language questions and await answers from other human users. Comparing to receiving results from web search engines using keyword queries, CQA users are likely to get more specific answers, because human answerers may catch the main point of the question. However, one of the key problems of this pattern is that sometimes no one helps to give answers, while web search engines hardly fail to response. In this paper, we analyze the not-answered questions and give a first try of predicting whether questions will receive answers. More specifically, we first analyze the questions of Yahoo Answers based on the features selected from different perspectives. Then, we formalize the prediction problem as supervised learning – binary classification problem and leverage the proposed features to make predictions. Extensive experiments are made on 76,251 questions collected from Yahoo! Answers. We analyze the specific characteristics of not-answered questions and try to suggest possible reasons for why a question is not likely to be answered. As for prediction, the experimental results show that classification based on the proposed features outperforms the simple word-based approach significantly.

【Keywords】:

203. Temporal Dynamics of User Interests in Tagging Systems.

Paper Link】 【Pages】:

【Authors】: Dawei Yin ; Liangjie Hong ; Zhenzhen Xue ; Brian D. Davison

【Abstract】: Collaborative tagging systems are now deployed extensivelyto help users share and organize resources.Tag prediction and recommendation systems generallymodel user behavior as research has shown that accuracycan be significantly improved by modeling users’preferences. However, these preferences are usuallytreated as constant over time, neglecting the temporalfactor within users’ interests. On the other hand, littleis known about how this factor may influence predictionin social bookmarking systems. In this paper, weinvestigate the temporal dynamics of user interests intagging systems and propose a user-tag-specific temporalinterests model for tracking users’ interests overtime. Additionally, we analyze the phenomenon of topicswitches in social bookmarking systems, showing that atemporal interests model can benefit from the integrationof topic switch detection and that temporal characteristicsof social tagging systems are different fromtraditional concept drift problems. We conduct experimentson three public datasets, demonstrating the importanceof personalization and user-tag specializationin tagging systems. Experimental results show that ourmethod can outperform state-of-the-art tag predictionalgorithms. We also incorporate our model within existingcontent-based methods yielding significant improvementsin performance.

【Keywords】:

204. Transfer Learning for Multiple-Domain Sentiment Analysis - Identifying Domain Dependent/Independent Word Polarity.

Paper Link】 【Pages】:

【Authors】: Yasuhisa Yoshida ; Tsutomu Hirao ; Tomoharu Iwata ; Masaaki Nagata ; Yuji Matsumoto

【Abstract】: Sentiment analysis is the task of determining the attitude (positive or negative) of documents. While the polarity of words in the documents is informative for this task, polarity of some words cannot be determined without domain knowledge. Detecting word polarity thus poses a challenge for multiple-domain sentiment analysis. Previous approaches tackle this problem with transfer learning techniques, but they cannot handle multiple source domains and multiple target domains. This paper proposes a novel Bayesian probabilistic model to handle multiple source and multiple target domains. In this model, each word is associated with three factors: Domain label, domain dependence/independence and word polarity. We derive an efficient algorithm using Gibbs sampling for inferring the parameters of the model, from both labeled and unlabeled texts. Using real data, we demonstrate the effectiveness of our model in a document polarity classification task compared with a method not considering the differences between domains. Moreover our method can also tell whether each word's polarity is domain-dependent or domain-independent. This feature allows us to construct a word polarity dictionary for each domain.

【Keywords】:

205. Propagating Both Trust and Distrust with Target Differentiation for Combating Web Spam.

Paper Link】 【Pages】:

【Authors】: Xianchao Zhang ; You Wang ; Nan Mou ; Wenxin Liang

【Abstract】: Propagating trust/distrust from a set of seed (good/bad) pages to the entire Web has been widely used to combat Web spam. It has been mentioned that a combined use of good and bad seeds can lead to better results. However, little work has been known to realize this insight successfully. A serious issue of existing algorithms is that trust/distrust is propagated in non-differential ways. However, it seems to be impossible to implement differential propagation if only trust or distrust is propagated. In this paper, we view that each Web page has both a trustworthy side and an untrustworthy side, and assign two scores to each Web page: T-Rank, scoring the trustworthiness, and D-Rank, scoring the untrustworthiness. We then propose an integrated framework which propagates both trust and distrust. In the framework, the propagation of T-Rank/D-Rank is penalized by the target's current D-Rank/T-Rank. In this way, propagating both trust and distrust with target differentiation is implemented. The proposed Trust-Distrust Rank (TDR) algorithm not only makes full use of both good seeds and bad seeds, but also overcomes the disadvantages of both existing trust propagation and distrust propagation algorithms. Experimental results show that TDR outperforms other typical anti-spam algorithms under various criteria.

【Keywords】:

206. Learning to Suggest Questions in Online Forums.

Paper Link】 【Pages】:

【Authors】: Tom Chao Zhou ; Chin-Yew Lin ; Irwin King ; Michael R. Lyu ; Young-In Song ; Yunbo Cao

【Abstract】: Online forums contain interactive and semantically related discussions on various questions. Extracted question-answer archive is invaluable knowledge, which can be used to improve Question Answering services. In this paper, we address the problem of Question Suggestion, which targets at suggesting questions that are semantically related to a queried question. Existing bag-of-words approaches suffer from the shortcoming that they could not bridge the lexical chasm between semantically related questions. Therefore, we present a new framework to suggest questions, and propose the Topicenhanced Translation-based Language Model (TopicTRLM) which fuses both the lexical and latent semantic knowledge. Extensive experiments have been conducted with a large real world data set. Experimental results indicate our approach is very effective and outperforms other popular methods in several metrics.

【Keywords】:

207. Heterogeneous Transfer Learning for Image Classification.

Paper Link】 【Pages】:

【Authors】: Yin Zhu ; Yuqiang Chen ; Zhongqi Lu ; Sinno Jialin Pan ; Gui-Rong Xue ; Yong Yu ; Qiang Yang

【Abstract】: Transfer learning as a new machine learning paradigm has gained increasing attention lately. In situations where the training data in a target domain are not sufficient to learn predictive models effectively, transfer learning leverages auxiliary source data from other related source domains for learning. While most of the existing works in this area only focused on using the source data with the same structure as the target data, in this paper, we push this boundary further by proposing a heterogeneous transfer learning framework for knowledge transfer between text and images. We observe that for a target-domain classification problem, some annotated images can be found on many social Web sites, which can serve as a bridge to transfer knowledge from the abundant text documents available over the Web. A key question is how to effectively transfer the knowledge in the source data even though the text can be arbitrarily found. Our solution is to enrich the representation of the target images with semantic concepts extracted from the auxiliary source data through a novel matrix factorization method. By using the latent semantic features generated by the auxiliary data, we are able to build a better integrated image classifier. We empirically demonstrate the effectiveness of our algorithm on the Caltech-256 image dataset.

【Keywords】:

Special Track on Computational Sustainability and AI 18

208. Green Driver: AI in a Microcosm.

Paper Link】 【Pages】:

【Authors】: Jim Apple ; Paul Chang ; Aran Clauson ; Heidi E. Dixon ; Hiba Fakhoury ; Matthew L. Ginsberg ; Erin Keenan ; Alex Leighton ; Kevin Scavezze ; Bryan Smith

【Abstract】: The Green Driver app is a dynamic routing application for GPS-enabled smartphones. Green Driver combines client GPS data with real-time traffic light information provided by cities to determine optimal routes in response to driver route requests. Routes are optimized with respect to travel time, with the intention of saving the driver both time and fuel, and rerouting can occur if warranted. During a routing session, client phones communicate with a centralized server that both collects GPS data and processes route requests. All relevant data are anonymized and saved to databases for analysis; statistics are calculated from the aggregate data and fed back to the routing engine to improve future routing. Analyses can also be performed to discern driver trends: where do drivers tend to go, how long do they stay, when and where does traffic congestion occur, and so on. The system uses a number of techniques from the field of artificial intelligence. We apply a variant of A* search for solving the stochastic shortest path problem in order to find optimal driving routes through a network of roads given light-status information. We also use dynamic programming and hidden Markov models to determine the progress of a driver through a network of roads from GPS data and light-status data. The Green Driver system is currently deployed for testing in Eugene, Oregon, and is scheduled for large-scale deployment in Portland, Oregon, in Spring 2011.

【Keywords】:

209. Enforcing Liveness in Autonomous Traffic Management.

Paper Link】 【Pages】:

【Authors】: Tsz-Chiu Au ; Neda Shahidi ; Peter Stone

【Abstract】: Looking ahead to the time when autonomous cars will be common, Dresner and Stone proposed a multiagent systems-based intersection control protocol called Autonomous Intersection Management (AIM). They showed that by leveraging the capacities of autonomous vehicles it is possible to dramatically reduce the time wasted in traffic, and therefore also fuel consumption and air pollution. The proposed protocol, however, handles reservation requests one at a time and does not prioritize reservations according to their relative priorities and waiting times, causing potentially large inequalities in granting reservations. For example, at an intersection between a main street and an alley, vehicles from the alley can take an excessively long time to get reservations to enter the intersection, causing a waste of time and fuel. The same is true in a network of intersections, in which gridlock may occur and cause traffic congestion. In this paper, we introduce the batch processing of reservations in AIM to enforce liveness properties in intersections and analyze the conditions under which no vehicle will get stuck in traffic. Our experimental results show that our prioritizing schemes outperform previous intersection control protocols in unbalanced traffic.

【Keywords】:

210. Policy Gradient Planning for Environmental Decision Making with Existing Simulators.

Paper Link】 【Pages】:

【Authors】: Mark Crowley ; David Poole

【Abstract】: In environmental and natural resource planning domains actions are taken at a large number of locations over multiple time periods. These problems have enormous state and action spaces, spatial correlation between actions, uncertainty and complex utility models. We present an approach for modeling these planning problems as factored Markov decision processes. The reward model can contain local and global components as well as spatial constraints between locations. The transition dynamics can be provided by existing simulators developed by domain experts. We propose a landscape policy defined as the equilibrium distribution of a Markov chain built from many locally-parameterized policies. This policy is optimized using a policy gradient algorithm. Experiments using a forestry simulator demonstrate the algorithm's ability to devise policies for sustainable harvest planning of a forest.

【Keywords】:

211. Dynamic Resource Allocation in Conservation Planning.

Paper Link】 【Pages】:

【Authors】: Daniel Golovin ; Andreas Krause ; Beth Gardner ; Sarah J. Converse ; Steve Morey

【Abstract】: Consider the problem of protecting endangered species by selecting patches of land to be used for conservation purposes. Typically, the availability of patches changes over time, and recommendations must be made dynamically. This is a challenging prototypical example of a sequential optimization problem under uncertainty in computational sustainability. Existing techniques do not scale to problems of realistic size. In this paper, we develop an efficient algorithm for adaptively making recommendations for dynamic conservation planning, and prove that it obtains near-optimal performance. We further evaluate our approach on a detailed reserve design case study of conservation planning for three rare species in the Pacific Northwest of the United States.

【Keywords】:

212. Water Conservation Through Facilitation on Residential Landscapes.

Paper Link】 【Pages】:

【Authors】: Rhonda Hoenigman ; Elizabeth Bradley ; Nichole Barger

【Abstract】: Plants can have positive effects on each other in numerous ways, including protection from harsh environmental conditions. This phenomenon, known as facilitation, occurs in water-stressed environments when shade from larger shrubs protects smaller annuals from harsh sun, enabling them to exist on scarce water. The topic of this paper is a model of this phenomenon that allows search algorithms to find residential landscape designs that incorporate facilitation to conserve water. This model is based in botany; it captures the growth requirements of real plant species in a fitness function, but also includes a penalty term in that function that encourages facilitative interactions with other plants on the landscape. To evaluate the effectiveness of this approach, two search strategies--simulated annealing and agent-based search--were applied to models of different collections of simulated plant types and landscapes with different light distributions. These two search strategies produced landscape designs with different spatial distributions of the larger plants. All designs exhibited facilitation and lower water use than designs where facilitation was not included.

【Keywords】:

213. Incorporating Boosted Regression Trees into Ecological Latent Variable Models.

Paper Link】 【Pages】:

【Authors】: Rebecca A. Hutchinson ; Li-Ping Liu ; Thomas G. Dietterich

【Abstract】: Important ecological phenomena are often observed indirectly. Consequently, probabilistic latent variable models provide an important tool, because they can include explicit models of the ecological phenomenon of interest and the process by which it is observed. However, existing latent variable methods rely on hand-formulated parametric models, which are expensive to design and require extensive preprocessing of the data. Nonparametric methods (such as regression trees) automate these decisions and produce highly accurate models. However, existing tree methods learn direct mappings from inputs to outputs — they cannot be applied to latent variable models. This paper describes a methodology for integrating nonparametric tree methods into probabilistic latent variable models by extending functional gradient boosting. The approach is presented in the context of occupancy-detection (OD) modeling, where the goal is to model the distribution of a species from imperfect detections. Experiments on 12 real and 3 synthetic bird species compare standard and tree-boosted OD models (latent variable models) with standard and tree-boosted logistic regression models (without latent structure). All methods perform similarly when predicting the observed variables, but the OD models learn better representations of the latent process. Most importantly, tree-boosted OD models learn the best latent representations when nonlinearities and interactions are present.

【Keywords】:

214. A Large-Scale Study on Predicting and Contextualizing Building Energy Usage.

Paper Link】 【Pages】:

【Authors】: J. Zico Kolter ; Joseph Ferreira Jr.

【Abstract】: In this paper we present a data-driven approach to modeling end user energy consumption in residential and commercial buildings. Our model is based upon a data set of monthly electricity and gas bills, collected by a utility over the course of several years, for approximately 6,500 buildings in Cambridge, MA. In addition, we use publicly available tax assessor records and geographical survey information to determine corresponding features for the buildings. Using both parametric and non-parametric learning methods, we learn models that predict distributions over energy usage based upon these features, and use these models to develop two end-user systems. For utilities or authorized institutions (those who may obtain access to the full data) we provide a system that visualizes energy consumption for each building in the city; this allows companies to quickly identify outliers (buildings which use much more energy than expected even after conditioning on the relevant predictors), for instance allowing them to target homes for potential retrofits or tiered pricing schemes. For other end users, we provide an interface for entering their own electricity and gas usage, along with basic information about their home, to determine how their consumption compares to that of similar buildings as predicted by our model. Merely allowing users to contextualize their consumption in this way, relating it to the consumption in similar buildings, can itself produce behavior changes to significantly reduce consumption.

【Keywords】:

215. The Steiner Multigraph Problem: Wildlife Corridor Design for Multiple Species.

Paper Link】 【Pages】:

【Authors】: Katherine J. Lai ; Carla P. Gomes ; Michael K. Schwartz ; Kevin S. McKelvey ; David E. Calkin ; Claire A. Montgomery

【Abstract】: The conservation of wildlife corridors between existing habitat preserves is important for combating the effects of habitat loss and fragmentation facing species of concern. We introduce the Steiner Multigraph Problem to model the problem of minimum-cost wildlife corridor design for multiple species with different landscape requirements. This problem can also model other analogous settings in wireless and social networks. As a generalization of Steiner forest, the goal is to find a minimum-cost subgraph that connects multiple sets of terminals. In contrast to Steiner forest, each set of terminals can only be connected via a subset of the nodes. Generalizing Steiner forest in this way makes the problem NP-hard even when restricted to two pairs of terminals. However, we show that if the node subsets have a nested structure, the problem admits a fixed-parameter tractable algorithm in the number of terminals. We successfully test exact and heuristic solution approaches on a wildlife corridor instance for wolverines and lynx in western Montana, showing that though the problem is computationally hard, heuristics perform well, and provably optimal solutions can still be obtained.

【Keywords】:

216. Hybrid Planning with Temporally Extended Goals for Sustainable Ocean Observing.

Paper Link】 【Pages】:

【Authors】: Hui Li ; Brian Williams

【Abstract】: A challenge to modeling and monitoring the health of the ocean environment is that it is largely under sensed and difficult to sense remotely. Autonomous underwater vehicles (AUVs) can improve observability, for example of algal bloom regions, ocean acidification, and ocean circulation. This AUV paradigm, however, requires robust operation that is cost effective and responsive to the environment. To achieve low cost we generate operational sequences automatically from science goals, and achieve robustness by reasoning about the discrete and continuous effects of actions. We introduce Kongming2, a generative planner for hybrid systems with temporally extended goals (TEGs) and temporally flexible actions. It takes as input high level goals and outputs trajectories and actions of the hybrid system, for example an AUV. Kongming2 makes two major extensions to Kongming1: planning for TEGs, and planning with temporally flexible actions. We demonstrated a proof of concept of the planner in the Atlantic ocean on Odyssey IV, an AUV designed and built by the MIT AUV Lab at Sea Grant.

【Keywords】:

217. Stochastic Model Predictive Controller for the Integration of Building Use and Temperature Regulation.

Paper Link】 【Pages】:

【Authors】: Alie El-Din Mady ; Gregory M. Provan ; Conor Ryan ; Kenneth N. Brown

【Abstract】: The aim of a modern Building Automation System (BAS) is to enhance interactive control strategies for energy efficiency and user comfort. In this context, we develop a novel control algorithm that uses a stochastic building occupancy model to improve mean energy efficiency while minimizing expected discomfort. We compare by simulation our Stochastic Model Predictive Control (SMPC) strategy to the standard heating control method to empirically demonstrate a 4.3% reduction in energy use and 38.3% reduction in expected discomfort.

【Keywords】:

218. Linear Dynamic Programs for Resource Management.

Paper Link】 【Pages】:

【Authors】: Marek Petrik ; Shlomo Zilberstein

【Abstract】: Sustainable resource management in many domains presents large continuous stochastic optimization problems, which can often be modeled as Markov decision processes (MDPs). To solve such large MDPs, we identify and leverage linearity in state and action sets that is common in resource management. In particular, we introduce linear dynamic programs (LDPs) that generalize resource management problems and partially observable MDPs (POMDPs). We show that the LDP framework makes it possible to adapt point-based methods--the state of the art in solving POMDPs--to solving LDPs. The experimental results demonstrate the efficiency of this approach in managing the water level of a river reservoir. Finally, we discuss the relationship with dual dynamic programming, a method used to optimize hydroelectric systems.

【Keywords】:

219. Logistic Methods for Resource Selection Functions and Presence-Only Species Distribution Models.

Paper Link】 【Pages】:

【Authors】: Steven Phillips ; Jane Elith

【Abstract】: In order to better protect and conserve biodiversity, ecologists use machine learning and statistics to understand how species respond to their environment and to predict how they will respond to future climate change, habitat loss and other threats. A fundamental modeling task is to estimate the probability that a given species is present in (or uses) a site, conditional on environmental variables such as precipitation and temperature. For a limited number of species, survey data consisting of both presence and absence records are available, and can be used to fit a variety of conventional classification and regression models. For most species, however, the available data consist only of occurrence records --- locations where the species has been observed. In two closely-related but separate bodies of ecological literature, diverse special-purpose models have been developed that contrast occurrence data with a random sample of available environmental conditions. The most widespread statistical approaches involve either fitting an exponential model of species' conditional probability of presence, or fitting a naive logistic model in which the random sample of available conditions is treated as absence data; both approaches have well-known drawbacks, and do not necessarily produce valid probabilities. After summarizing existing methods, we overcome their drawbacks by introducing a new scaled binomial loss function for estimating an underlying logistic model of species presence/absence. Like the Expectation-Maximization approach of Ward et al. and the method of Steinberg and Cardell, our approach requires an estimate of population prevalence, $\Pr(y=1)$, since prevalence is not identifiable from occurrence data alone. In contrast to the latter two methods, our loss function is straightforward to integrate into a variety of existing modeling frameworks such as generalized linear and additive models and boosted regression trees. We also demonstrate that approaches by Lele and Keim and by Lancaster and Imbens that surmount the identifiability issue by making parametric data assumptions do not typically produce valid probability estimates.

【Keywords】:

220. Modeling and Monitoring Crop Disease in Developing Countries.

Paper Link】 【Pages】:

【Authors】: John Alexander Quinn ; Kevin Leyton-Brown ; Ernest Mwebaze

【Abstract】: Information about the spread of crop disease is vital in developing countries, and as a result the governments of such countries devote scarce resources to gathering such data. Unfortunately, current surveys tend to be slow and expensive, and hence also tend to gather insufficient quantities of data. In this work we describe three general methods for improving the use of survey resources by performing data collection with mobile devices and by directing survey progress through the application of AI techniques. First, we describe a spatial disease density model based on Gaussian process ordinal regression, which offers a better representation of the disease level distribution, as compared to the statistical approaches typically applied. Second, we show how this model can be used to dynamically route survey teams to obtain the most valuable survey possible given a fixed budget. Third, we demonstrate that the diagnosis of plant disease can be automated using images taken by a camera phone, enabling data collection by survey workers with only basic training. We have applied our methods to the specific challenge of viral cassava disease monitoring in Uganda, for which we have implemented a real-time mobile survey system that will soon see practical use.

【Keywords】:

221. Learned Behaviors of Multiple Autonomous Agents in Smart Grid Markets.

Paper Link】 【Pages】:

【Authors】: Prashant P. Reddy ; Manuela M. Veloso

【Abstract】: One proposed approach to managing a large complex Smart Grid is through Broker Agents who buy electrical power from distributed producers, and also sell power to consumers, via a Tariff Market--a new market mechanism where Broker Agents publish concurrent bid and ask prices. A key challenge is the specification of the market strategy that the Broker Agents should use in order to earn profits while maintaining the market's balance of supply and demand. Interestingly, previous work has shown that a Broker Agent can learn its strategy, using Markov Decision Processes (MDPs) and Q-learning, and outperform other Broker Agents that use predetermined or randomized strategies. In this work, we investigate the more representative scenario in which multiple Broker Agents, instead of a single one, are independently learning their strategies. Using a simulation environment based on real data, we find that Broker Agents who employ periodic increases in exploration achieve higher rewards. We also find that varying levels of market dominance in customer allocation models result in remarkably distinct outcomes in market prices and aggregate Broker Agent rewards. The latter set of results can be explained by established economic principles regarding the emergence of monopolies in market-based competition, further validating our approach.

【Keywords】:

222. Efficient Energy-Optimal Routing for Electric Vehicles.

Paper Link】 【Pages】:

【Authors】: Martin Sachenbacher ; Martin Leucker ; Andreas Artmeier ; Julian Haselmayr

【Abstract】: Traditionally routing has focused on finding shortest paths in networks with positive, static edge costs representing the distance between two nodes. Energy-optimal routing for electric vehicles creates novel algorithmic challenges, as simply understanding edge costs as energy values and applying standard algorithms does not work. First, edge costs can be negative due to recuperation, excluding Dijkstra-like algorithms. Second, edge costs may depend on parameters such as vehicle weight only known at query time, ruling out existing preprocessing techniques. Third, considering battery capacity limitations implies that the cost of a path is no longer just the sum of its edge costs. This paper shows how these challenges can be met within the framework of A* search. We show how the specific domain gives rise to a consistent heuristic function yielding an O(n 2 ) routing algorithm. Moreover, we show how battery constraints can be treated by dynamically adapting edge costs and hence can be handled in the same way as parameters given at query time, without increasing run-time complexity. Experimental results with real road networks and vehicle data demonstrate the advantages of our solution.

【Keywords】:

223. Verifying Intervention Policies to Counter Infection Propagation over Networks: A Model Checking Approach.

Paper Link】 【Pages】:

【Authors】: Ganesh Ram Santhanam ; Yuly Suvorov ; Samik Basu ; Vasant Honavar

【Abstract】: Spread of infections (diseases, ideas, etc.) in a network can be modeled as the evolution of states of nodes in a graph as a function of the states of their neighbors. Given an initial configuration of a network in which a subset of the nodes have been infected, and an infection propagation function that specifies how the states of the nodes evolve over time, we show how to use model checking to identify, verify, and evaluate the effectiveness of intervention policies for containing the propagation of infection over such networks.

【Keywords】:

224. Discovering Life Cycle Assessment Trees from Impact Factor Databases.

Paper Link】 【Pages】:

【Authors】: Naren Sundaravaradan ; Debprakash Patnaik ; Naren Ramakrishnan ; Manish Marwah ; Amip Shah

【Abstract】: In recent years, environmental sustainability has received widespread attention due to continued depletion of natural resources and degradation of the environment. Life cycle assessment (LCA) is a methodology for quantifying multiple environmental impacts of a product, across its entire life cycle — from creation to use to discard. The key object of interest in LCA is the inventory tree, with the desired product as the root node and the materials and processes used across its life cycle as the children. The total impact of the parent in any environmental category is a linear combination of the impacts of the children in that category. LCA has generally been used in "forward: mode: given an inventory tree and impact factors of its children, the task is to compute the impact factors of the root, i.e., the product being modeled. We propose a data mining approach to solve the inverse problem, where the task is to infer inventory trees from a database of environmental factors. This is an important problem with applications in not just understanding what parts and processes constitute a product but also in designing and developing more sustainable alternatives. Our solution methodology is one of feature selection but set in the context of a non-negative least squares problem. It organizes numerous non-negative least squares fits over the impact factor database into a set of pairwise membership relations which are then summarized into candidate trees in turn yielding a consensus tree. We demonstrate the applicability of our approach over real LCA datasets obtained from a large computer manufacturer.

【Keywords】:

225. Decentralised Control of Micro-Storage in the Smart Grid.

Paper Link】 【Pages】:

【Authors】: Thomas Voice ; Perukrishnen Vytelingum ; Sarvapali D. Ramchurn ; Alex Rogers ; Nicholas R. Jennings

【Abstract】: In this paper, we propose a novel decentralised control mechanism to manage micro-storage in the smart grid. Our approach uses an adaptive pricing scheme that energy suppliers apply to home smart agents controlling micro-storage devices. In particular, we prove that the interaction between a supplier using our pricing scheme and the actions of selfish micro-storage agents forms a globally stable feedback loop that converges to an efficient equilibrium. We further propose a market strategy that allows the supplier to reduce wholesale purchasing costs without increasing the uncertainty and variance for its aggregate consumer demand. Moreover, we empirically evaluate our mechanism (based on the UK grid data) and show that it yields savings of up to 16% in energy cost for consumers using storage devices with average capacity 10 kWh. Furthermore, we show that it is robust against extreme system changes.

【Keywords】:

Special Track on Integrated Intelligence 5

226. Analogical Dialogue Acts: Supporting Learning by Reading Analogies in Instructional Texts.

Paper Link】 【Pages】:

【Authors】: David Michael Barbella ; Kenneth D. Forbus

【Abstract】: Analogy is heavily used in instructional texts. We introduce the concept of analogical dialogue acts (ADAs), which represent the roles utterances play in instructional analogies. We describe a catalog of such acts, based on ideas from structure-mapping theory. We focus on the operations that these acts lead to while understanding instructional texts, using the Structure-Mapping Engine (SME) and dynamic case construction in a computational model. We test this model on a small corpus of instructional analogies expressed in simplified English, which were understood via a semi-automatic natural language system using analogical dialogue acts. The model enabled a system to answer questions after understanding the analogies that it was not able to answer without them.

【Keywords】:

227. Cognitive Synergy between Procedural and Declarative Learning in the Control of Animated and Robotic Agents Using the OpenCogPrime AGI Architecture.

Paper Link】 【Pages】:

【Authors】: Ben Goertzel ; Joel Pitt ; Jared Wigmore ; Nil Geisweiller ; Zhenhua Cai ; Ruiting Lian ; Deheng Huang ; Gino Yu

【Abstract】: The hypothesis is presented that "cognitive synergy" -- proactive and mutually-assistive feedback between different cognitive processes associated with different types of memory -- may serve as a foundation for advanced artificial general intelligence. A specific AI architecture founded on this idea, OpenCogPrime, is described, in the context of its application to control virtual agents and robots. The manifestations of cognitive synergy in OpenCogPrime's procedural and declarative learning algorithms are discussed in some detail.

【Keywords】:

228. Contextually-Based Utility: An Appraisal-Based Approach at Modeling Framing and Decisions.

Paper Link】 【Pages】:

【Authors】: Jonathan Yasuo Ito ; Stacy Marsella

【Abstract】: Creating accurate computational models of human decision making is a vital step towards the realization of socially intelligent systems capable of both predicting and simulating human behavior. In modeling human decision making, a key factor is the psychological phenomenon known as "framing", in which the preferences of a decision maker change in response to contextual changes in decision problems. Existing approaches treat framing as a one-dimensional contextual influence based on the perception of outcomes as either gains or losses. However, empirical studies have shown that framing effects are much more multifaceted than one-dimensional views of framing suggest. To address this limitation, we propose an integrative approach to modeling framing which combines the psychological principles of cognitive appraisal theories and decision-theoretic notions of utility and probability. We show that this approach allows for both the identification and computation of the salient contextual factors in a decision as well as modeling how they ultimately affect the decision process. Furthermore, we show that our multi-dimensional, appraisal-based approach can account for framing effects identified in the empirical literature which cannot be addressed by one-dimensional theories, thereby promising more accurate models of human behavior.

【Keywords】:

229. Combining Learned Discrete and Continuous Action Models.

Paper Link】 【Pages】:

【Authors】: Joseph Z. Xu ; John E. Laird

【Abstract】: Action modeling is an important skill for agents that must perform tasks in novel domains. Previous work on action modeling has focused on learning STRIPS operators in discrete, relational domains. There has also been a separate vein of work in continuous function approximation for use in optimal control in robotics. Most real world domains are grounded in continuous dynamics but also exhibit emergent regularities at an abstract relational level of description. These two levels of regularity are often difficult to capture using a single action representation and learning method. In this paper we describe a system that combines discrete and continuous action modeling techniques in the Soar cognitive architecture. Our system accepts a continuous state representation from the environment and derives a relational state on top of it using spatial relations. The dynamics over each representation is learned separately using two simple instance-based algorithms. The predictions from the individual models are then combined in a way that takes advantage of the information captured by each representation. We empirically show that this combined model is more accurate and generalizable than each of the individual models in a spatial navigation domain.

【Keywords】:

230. Cross Media Entity Extraction and Linkage for Chemical Documents.

Paper Link】 【Pages】:

【Authors】: Su Yan ; W. Scott Spangler ; Ying Chen

【Abstract】: Text and images are two major sources of information in scientific literature. Information from these two media typically reinforce and complement each other, thus simplifying the process for human to extract and comprehend information. However, machines cannot create the links or have the semantic understanding between images and text. We propose to integrate text analysis and image processing techniques to bridge the gap between the two media, and discover knowledge from the combined information sources, which would be otherwise lost by traditional single-media based mining systems. The focus is on the chemical entity extraction task because images are well known to add value to the textual content in chemical literature. Annotation of US chemical patent documents demonstrates the effectiveness of our proposal.

【Keywords】:

Special Track on New Scientific and Technical Advances in Research 12

231. Effective End-User Interaction with Machine Learning.

Paper Link】 【Pages】:

【Authors】: Saleema Amershi ; James Fogarty ; Ashish Kapoor ; Desney S. Tan

【Abstract】: End-user interactive machine learning is a promising tool for enhancing human productivity and capabilities with large unstructured data sets. Recent work has shown that we can create end-user interactive machine learning systems for specific applications. However, we still lack a generalized understanding of how to design effective end-user interaction with interactive machine learning systems. This work presents three explorations in designing for effective end-user interaction with machine learning in CueFlik, a system developed to support Web image search. These explorations demonstrate that interactions designed to balance the needs of end-users and machine learning algorithms can significantly improve the effectiveness of end-user interactive machine learning.

【Keywords】:

232. Global Seismic Monitoring: A Bayesian Approach.

Paper Link】 【Pages】:

【Authors】: Nimar S. Arora ; Stuart J. Russell ; Paul Kidwell ; Erik B. Sudderth

【Abstract】: The automated processing of multiple seismic signals to detect and localize seismic events is a central tool in both geophysics and nuclear treaty verification. This paper reports on a project, begun in 2009, to reformulate this problem in a Bayesian framework. A Bayesian seismic monitoring system, NET-VISA, has been built comprising a spatial event prior and generative models of event transmission and detection, as well as an inference algorithm. Applied in the context of the International Monitoring System (IMS), a global sensor network developed for the Comprehensive Nuclear-Test-Ban Treaty (CTBT), NET-VISA achieves a reduction of around 50% in the number of missed events compared to the currently deployed system. It also finds events that are missed even by the human analysts who post-process the IMS output.

【Keywords】:

233. The Next Best Solution.

Paper Link】 【Pages】:

【Authors】: Ronen I. Brafman ; Enrico Pilotto ; Francesca Rossi ; Domenico Salvagnin ; Kristen Brent Venable ; Toby Walsh

【Abstract】: We study the computational complexity of finding the next most preferred solution in some common formalisms for representing constraints and preferences. The problem is computationally intractable for CSPs, but is polynomial for tree-shaped CSPs and tree-shaped fuzzy CSPs. On the other hand, it is intractable for weighted CSPs, even under restrictions on the constraint graph. For CP-nets, the problem is polynomial when the CP-net is acyclic. This remains so if we add (soft) constraints that are tree-shaped and topologically compatible with the CP-net.

【Keywords】:

234. New Expressive Languages for Ontological Query Answering.

Paper Link】 【Pages】:

【Authors】: Andrea Calì ; Georg Gottlob ; Andreas Pieris

【Abstract】: Ontology-based data access is a powerful form of extending database technology, where a classical extensional database (EDB) is enhanced by an ontology that generates new intensional knowledge which may contribute to answer a query. Recently, the Datalog+/- family of ontology languages was introduced; in Datalog+/-, rules are tuple-generating dependencies (TGDs), i.e., Datalog rules with the possibility of having existentially-quantified variables in the head. In this paper we introduce a novel Datalog+/- language, namely sticky sets of TGDs, which allows for a wide class of joins in the body, while enjoying at the same time a low query-answering complexity. We establish complexity results for answering conjunctive queries under sticky sets of TGDs, showing, in particular, that ontological conjunctive queries can be compiled into first-order and thus SQL queries over the given EDB instance. We also show some extensions of sticky sets of TGDs, and how functional dependencies and so-called negative constraints can be added to a sticky set of TGDs without increasing the complexity of query answering. Our language thus properly generalizes both classical database constraints and most widespread tractable description logics.

【Keywords】:

235. Quantity Makes Quality: Learning with Partial Views.

Paper Link】 【Pages】:

【Authors】: Nicolò Cesa-Bianchi ; Shai Shalev-Shwartz ; Ohad Shamir

【Abstract】: In many real world applications, the number of examples to learn from is plentiful, but we can only obtain limited information on each individual example. We study the possibilities of efficient, provably correct, large-scale learning in such settings. The main theme we would like to establish is that large amounts of examples can compensate for the lack of full information on each individual example. The type of partial information we consider can be due to inherent noise or from constraints on the type of interaction with the data source. In particular, we describe and analyze algorithms for budgeted learning, in which the learner can only view a few attributes of each training example, and algorithms for learning kernel-based predictors, when individual examples are corrupted by random noise.

【Keywords】:

236. Design and Analysis of Value Creation Networks.

Paper Link】 【Pages】:

【Authors】: Sampath Kameshwaran ; Sameep Mehta ; Vinayaka Pandit

【Abstract】: There are many diverse domains like academic collaboration, service industry, and movies, where a group of agents are involved in a set of activities through interactions or collaborations to create value. The end result of the value creation process is two pronged: firstly, there is a cumulative value created due to the interactions and secondly, a network that captures the pattern of historical interactions between the agents. In this paper we summarize our efforts towards design and analysis of value creation networks: 1) network representation of interactions and value creations, 2) identify contribution of a node based on values created from various activities, and 3) ranking nodes based on structural properties of interactions and the resulting values. To highlight the efficacy of our proposed algorithms, we present results on IMDB and services industry data.

【Keywords】:

237. Two Visual Strategies for Solving the Raven's Progressive Matrices Intelligence Test.

Paper Link】 【Pages】:

【Authors】: Maithilee Kunda ; Keith McGreggor ; Ashok K. Goel

【Abstract】: We present two visual algorithms, called the affine and fractal methods, which each solve a considerable portion of the Raven’s Progressive Matrices (RPM) test. The RPM is considered to be one of the premier psychometric measures of general intelligence. Current computational accounts of the RPM assume that visual test inputs are translated into propositional representations before further reasoning takes place. We propose that visual strategies can also solve RPM problems, in line with behavioral evidence showing that humans do use visual strategies to some extent on the RPM. Our two visual methods currently solve RPM problems at the level of typical 9- to 10-year-olds.

【Keywords】:

238. A POMDP-Based Optimal Control of P300-Based Brain-Computer Interfaces.

Paper Link】 【Pages】:

【Authors】: Jaeyoung Park ; Kee-Eung Kim ; Yoon-Kyu Song

【Abstract】: Most of the previous work on brain-computer interfaces (BCIs) exploiting the P300 in electroencephalography (EEG) has focused on low-level signal processing algorithms such as feature extraction and classification methods. Although a significant improvement has been made in the past, the accuracy of detecting P300 is limited by the inherently low signal-to-noise ratio in EEGs. In this paper, we present a systematic approach to optimize the interface using partially observable Markov decision processes (POMDPs). Through experiments involving human subjects, we show the P300 speller system that is optimized using the POMDP achieves a significant performance improvement in terms of the communication bandwidth in the interaction.

【Keywords】:

239. Planning with Specialized SAT Solvers.

Paper Link】 【Pages】:

【Authors】: Jussi Rintanen

【Abstract】: Logic, and declarative representation of knowledge in general, have long been a preferred framework for problem solving in AI. However, specific subareas of AI have been eager to abandon general-purpose knowledge representation in favor of methods that seem to address their computational core problems better. In planning, for example, state-space search has in the last several years been preferred to logic-based methods such as SAT. In our recent work, we have demonstrated that the observed performance differences between SAT and specialized state-space search methods largely go back to the difference between a blind (or at least planning-agnostic) and a planning-specific search method. If SAT search methods are given even simple heuristics which make the search goal-directed, the efficiency differences disappear.

【Keywords】:

240. Termination and Correctness Analysis of Cyclic Control.

Paper Link】 【Pages】:

【Authors】: Siddharth Srivastava ; Neil Immerman ; Shlomo Zilberstein

【Abstract】: The utility of including cyclic flows of control in plans has been long recognized by the planning community. Loops in a plan help increase both its applicability and the compactness of representation. However, progress in finding such plans has been limited largely due to lack of methods for reasoning about the correctness and safety properties of loops of actions. We present an overview of recent results for determining the class of problems that a plan with loops can solve. These methods can be used to direct the construction of a rich new form of generalized plans that solve a desired class of problems.

【Keywords】:

241. Recommendation Sets and Choice Queries: There Is No Exploration/Exploitation Tradeoff!

Paper Link】 【Pages】:

【Authors】: Paolo Viappiani ; Craig Boutilier

【Abstract】: Utility elicitation is an important component of many applications, such as decision support systems and recommender systems. Such systems query users about their preferences and offer recommendations based on the system's belief about the user's utility function. We analyze the connection between the problem of generating optimal recommendation sets and the problem of generating optimal choice queries, considering both Bayesian and regret-based elicitation. Our results show that, somewhat surprisingly, under very general circumstances, the optimal recommendation set coincides with the optimal query.

【Keywords】:

242. End-User Feature Labeling via Locally Weighted Logistic Regression.

Paper Link】 【Pages】:

【Authors】: Weng-Keen Wong ; Ian Oberst ; Shubhomoy Das ; Travis Moore ; Simone Stumpf ; Kevin McIntosh ; Margaret M. Burnett

【Abstract】: Applications that adapt to a particular end user often make inaccurate predictions during the early stages when training data is limited. Although an end user can improve the learning algorithm by labeling more training data, this process is time consuming and too ad hoc to target a particular area of inaccuracy. To solve this problem, we propose a new learning algorithm based on Locally Weighted Logistic Regression for feature labeling by end users, enabling them to point out which features are important for a class, rather than provide new training instances. In our user study, the first allowing ordinary end users to freely choose features to label directly from text documents, our algorithm was more effective than others at leveraging end users’ feature labels to improve the learning algorithm. Our results strongly suggest that allowing users to freely choose features to label is a promising method for allowing end users to improve learning algorithms effectively.

【Keywords】:

Special Track on Physically Grounded AI 10

243. Multi-Observation Sensor Resetting Localization with Ambiguous Landmarks.

Paper Link】 【Pages】:

【Authors】: Brian Coltin ; Manuela M. Veloso

【Abstract】: Successful approaches to the robot localization problem include Monte Carlo particle filters, which estimate non-parametric localization belief distributions. However, particle filters fare poorly at determining the robot's position without a good initial hypothesis. This problem has been addressed for robots that sense visual landmarks with sensor resetting, by performing sensor-based resampling when the robot is lost. For robots that make sparse, ambiguous and noisy observations, standard sensor resetting places new location hypotheses across a wide region, in positions that may be inconsistent with previous observations. We propose Multi-Observation Sensor Resetting, where observations from multiple frames are merged to generate new hypotheses more effectively. We demonstrate experimentally in the robot soccer domain on the NAO humanoid robots that Multi-Observation Sensor Resetting converges more efficiently to the robot's true position than standard sensor resetting, and is more robust to systematic vision errors.

【Keywords】:

244. Autonomous Skill Acquisition on a Mobile Manipulator.

Paper Link】 【Pages】:

【Authors】: George Konidaris ; Scott Kuindersma ; Roderic A. Grupen ; Andrew G. Barto

【Abstract】: We describe a robot system that autonomously acquires skills through interaction with its environment. The robot learns to sequence the execution of a set of innate controllers to solve a task, extracts and retains components of that solution as portable skills, and then transfers those skills to reduce the time required to learn to solve a second task.

【Keywords】:

245. A Scalable Tree-Based Approach for Joint Object and Pose Recognition.

Paper Link】 【Pages】:

【Authors】: Kevin Lai ; Liefeng Bo ; Xiaofeng Ren ; Dieter Fox

【Abstract】: Recognizing possibly thousands of objects is a crucial capability for an autonomous agent to understand and interact with everyday environments. Practical object recognition comes in multiple forms: Is this a coffee mug (category recognition). Is this Alice's coffee mug? (instance recognition). Is the mug with the handle facing left or right? (pose recognition). We present a scalable framework, Object-Pose Tree, which efficiently organizes data into a semantically structured tree. The tree structure enables both scalable training and testing, allowing us to solve recognition over thousands of object poses in near real-time. Moreover, by simultaneously optimizing all three tasks, our approach outperforms standard nearest neighbor and 1-vs-all classifications, with large improvements on pose recognition. We evaluate the proposed technique on a dataset of 300 household objects collected using a Kinect-style 3D camera. Experiments demonstrate that our system achieves robust and efficient object category, instance, and pose recognition on challenging everyday objects.

【Keywords】:

246. Recognizing Text Through Sound Alone.

Paper Link】 【Pages】:

【Authors】: Wenzhe Li ; Tracy Anne Hammond

【Abstract】: This paper presents an acoustic sound recognizer to recognize what people are writing on a table or wall by utilizing the sound signal information generated from a key, pen, or fingernail moving along a textured surface. Sketching provides a natural modality to interact with text, and sound is an effective modality for distinguishing text. However, limited research has been conducted in this area. Our system uses a dynamic time- warping approach to recognize 26 hand-sketched characters (A-Z) solely through their acoustic signal. Our initial prototype system is user-dependent and relies on fixed stroke ordering. Our algorithm relied mainly on two features: mean amplitude and MFCCs (Mel-frequency cepstral coefficients). Our results showed over 80% recognition accuracy.

【Keywords】:

247. DISCO: Describing Images Using Scene Contexts and Objects.

Paper Link】 【Pages】:

【Authors】: Ifeoma Nwogu ; Yingbo Zhou ; Christopher Brown

【Abstract】: In this paper, we propose a bottom-up approach to generating short descriptive sentences from images, to enhance scene understanding. We demonstrate automatic methods for mapping the visual content in an image to natural spoken or written language. We also introduce a human-in-the-loop evaluation strategy that quantitatively captures the meaningfulness of the generated sentences. We recorded a correctness rate of 60.34% when human users were asked to judge the meaningfulness of the sentences generated from relatively challenging images. Also, our automatic methods compared well with the state-of-the-art techniques for the related computer vision tasks.

【Keywords】:

248. Continuous Occupancy Mapping with Integral Kernels.

Paper Link】 【Pages】:

【Authors】: Simon Timothy O'Callaghan ; Fabio T. Ramos

【Abstract】: We address the problem of building a continuous occupancy representation of the environment with ranging sensors. Observations from such sensors provide two types of information: a line segment or a beam indicating no returns along them (free-space); a point or return at the end of the segment representing an occupied surface. To model these two types of observations in a principled statistical manner, we propose a novel methodology based on integral kernels. We show that integral kernels can be directly incorporated into a Gaussian process classification (GPC) framework to provide a continuous non-parametric Bayesian estimation of occupancy. Directly handling line segment and point observations avoids the need to discretise segments into points, reducing the computational cost of GPC inference and learning. We present experiments on 2D and 3D datasets demonstrating the benefits of the approach.

【Keywords】:

249. Learning Accuracy and Availability of Humans Who Help Mobile Robots.

Paper Link】 【Pages】:

【Authors】: Stephanie Rosenthal ; Manuela M. Veloso ; Anind K. Dey

【Abstract】: When mobile robots perform tasks in environments with humans, it seems appropriate for the robots to rely on such humans for help instead of dedicated human oracles or supervisors. However, these humans are not always available nor always accurate. In this work, we consider human help to a robot as concretely providing observations about the robot's state to reduce state uncertainty as it executes its policy autonomously. We model the probability of receiving an observation from a human in terms of their availability and accuracy by introducing Human Observation Providers POMDPs (HOP-POMDPs). We contribute an algorithm to learn human availability and accuracy online while the robot is executing its current task policy. We demonstrate that our algorithmis effective in approximating the true availability and accuracy of humans without depending on oracles to learn, thus increasing the tractability of deploying a robot that can occasionally ask for help.

【Keywords】:

Paper Link】 【Pages】:

【Authors】: Stefanie Tellex ; Thomas Kollar ; Steven Dickerson ; Matthew R. Walter ; Ashis Gopal Banerjee ; Seth J. Teller ; Nicholas Roy

【Abstract】: This paper describes a new model for understanding natural language commands given to autonomous systems that perform navigation and mobile manipulation in semi-structured environments. Previous approaches have used models with fixed structure to infer the likelihood of a sequence of actions given the environment and the command. In contrast, our framework, called Generalized Grounding Graphs, dynamically instantiates a probabilistic graphical model for a particular natural language command according to the command's hierarchical and compositional semantic structure. Our system performs inference in the model to successfully find and execute plans corresponding to natural language commands such as "Put the tire pallet on the truck." The model is trained using a corpus of commands collected using crowdsourcing. We pair each command with robot actions and use the corpus to learn the parameters of the model. We evaluate the robot's performance by inferring plans from natural language commands, executing each plan in a realistic robot simulator, and asking users to evaluate the system's performance. We demonstrate that our system can successfully follow many natural language commands from the corpus.

【Keywords】:

251. Balancing Safety and Exploitability in Opponent Modeling.

Paper Link】 【Pages】:

【Authors】: Zhikun Wang ; Abdeslam Boularias ; Katharina Mülling ; Jan Peters

【Abstract】: Opponent modeling is a critical mechanism in repeated games. It allows a player to adapt its strategy in order to better respond to the presumed preferences of his opponents. We introduce a new modeling technique that adaptively balances exploitability and risk reduction. An opponent’s strategy is modeled with a set of possible strategies that contain the actual strategy with a high probability. The algorithm is safe as the expected payoff is above the minimax payoff with a high probability, and can exploit the opponents’ preferences when sufficient observations have been obtained. We apply them to normal-form games and stochastic games with a finite number of stages. The performance of the proposed approach is first demonstrated on repeated rock-paper-scissors games. Subsequently, the approach is evaluated in a human-robot table-tennis setting where the robot player learns to prepare to return a served ball. By modeling the human players, the robot chooses a forehand, backhand or middle preparation pose before they serve. The learned strategies can exploit the opponent’s preferences, leading to a higher rate of successful returns.

【Keywords】:

252. Self-Aware Traffic Route Planning.

Paper Link】 【Pages】:

【Authors】: David Wilkie ; Jur P. van den Berg ; Ming C. Lin ; Dinesh Manocha

【Abstract】: One of the most ubiquitous AI applications is vehicle route planning. While state-of-the-art systems take into account current traffic conditions or historic traffic data, current planning approaches ignore the impact of their own plans on the future traffic conditions. We present a novel algorithm for self-aware route planning that uses the routes it plans for current vehicle traffic to more accurately predict future traffic conditions for subsequent cars. Our planner uses a roadmap with stochastic, time-varying traffic densities that are defined by a combination of historical data and the densities predicted by the planned routes for the cars ahead of the current traffic. We have applied our algorithm to large-scale traffic route planning, and demonstrated that our self-aware route planner can more accurately predict future traffic conditions, which results in a reduction of the travel time for those vehicles that use our algorithm.

【Keywords】:

AAAI / SIGART Doctoral Consortium 15

253. The AC(C) Language: Integrating Answer Set Programming and Constraint Logic Programming.

Paper Link】 【Pages】:

【Authors】: Forrest Sheng Bao

【Abstract】: Combining Answer Set Programming (ASP) and Constraint Logic Programming (CLP) can create a more powerful language for knowledge representation and reasoning. The language AC(C) is designed to integrate ASP and CLP. Compared with existing integration of ASP and CSP, AC(C) allows representing user-defined constraints. Such integration provides great power for applications requiring logical reasoning involving constraints, e.g., temporal planning. In AC(C), user-defined and primitive constraints can be solved by a CLP inference engine while the logical reasoning over those constraints and regular logic literals is solved by an ASP inference engine (i.e., solver). My PhD work includes improving the language AC(C), implementing its faster inference engine and investigating how effective the new system can be used to solve a challenging application, temporal planning.

【Keywords】:

254. Joint Inference for Extracting Text Descriptors from Triage Images of Mass Disaster Victims.

Paper Link】 【Pages】:

【Authors】: Niyati Chhaya

【Abstract】: The major contributions of this work include a set of biographical feature extractors brought together by a probabilistic graphical model. The graphical model acts as a communication medium between different feature extractors, together resulting in a text descriptor to describe triage images of disaster victims. The model is built using domain information gathered from data and literature.

【Keywords】:

255. Model Update for Automated Planning.

Paper Link】 【Pages】:

【Authors】: Maria Viviane de Menezes ; Leliane Nunes de Barros

【Abstract】: Model update is a formal approach to correct a system model M w.r.t some property not satisfied by M. In this work, we show how this formal approach can be used for plan and planning domain verification and update. While a model checking method can directly be used to perform plan verification, model update techniques can be used to either update an incorrect plan and\or update a planning domain specification. Well known model update approaches are based on CTL — a logic which does not take into account the actions. In previous work, we have proposed the alpha-CTL logic, a logic whose semantics is based on actions. Here, we are proposing a model update system based on alpha-CTL which is able to automatically modify a plan M, generating a new plan M' that satisfies phi or, if there is not such a plan, to automatically update the corresponding planning domain.

【Keywords】:

256. Long-Term Declarative Memory for Generally Intelligent Agents.

Paper Link】 【Pages】:

【Authors】: Nate Derbinsky

【Abstract】: The goal of my research is to develop and evaluate long-term declarative memory mechanisms that are effective and efficient across a variety of tasks.

【Keywords】:

257. Ensemble Classification for Relational Domains.

Paper Link】 【Pages】:

【Authors】: Hoda Eldardiry

【Abstract】: Ensemble classification methods have been shown to produce more accurate predictions than the base component models. Due to their effectiveness, ensemble approaches have been applied in a wide range of domains to improve classification. The expected prediction error of classification models can be decomposed into bias and variance. Ensemble methods that independently construct component models (e.g., bagging) can improve performance by reducing the error due to variance, while methods that dependently construct component models (e.g., boosting) can improve performance by reducing the error due to bias and variance. Although ensemble methods were initially developed for classification of independent and identically distributed (i.i.d.) data, they can be directly applied for relational data by using a relational classifier as the base component model. This straightforward approach can improve classification for network data, but suffers from a number of limitations. First, relational data characteristics will only be exploited by the base relational classifier, and not by the ensemble algorithm itself. We note that explicitly accounting for the structured nature of relational data by the ensemble mechanism can significantly improve ensemble classification. Second, ensemble learning methods that assume i.i.d. data can fail to preserve the relational structure of non-i.i.d. data, which will (1) prevent the relational base classifiers from exploiting these structures, and (2) fail to accurately capture properties of the dataset, which can lead to inaccurate models and classifications. Third, ensemble mechanisms that assume i.i.d. data are limited to reducing errors associated with i.i.d. models and fail to reduce additional sources of error associated with more powerful (e.g., collective classification models. Our key observation is that collective classification methods have error due to variance in inference. This has been overlooked by current ensemble methods that assume exact inference methods and only focus on the typical goal of reducing errors due to learning, even if the methods explicitly consider relational data. Here we study the problem of ensemble classification for relational domains by focusing on the reduction of error due to variance. We propose a relational ensemble framework that explicitly accounts for the structured nature of relational data during both learning and inference. Our proposed framework consists of two components. (1) A method for learning accurate ensembles from relational data, focusing on the reduction of error due to variance in learning, while preserving the relational characteristics in the data. (2) A method for applying ensembles in collective classification contexts, focusing on further reduction of the error due to variance in inference, which has not been considered in state of the art ensemble methods.

【Keywords】:

258. Developing a Language for Spoken Programming.

Paper Link】 【Pages】:

【Authors】: Benjamin M. Gordon

【Abstract】: The dominant paradigm for programming a computer today is text entry via keyboard and mouse, but there aremany common situations where this is not ideal. I address this through the creation of a new language thatis explicitly intended for spoken programming. In addition, I describe a supporting editor that improvesrecognition accuracy by making use of type information and scoping to increase recognizer context.

【Keywords】:

259. A Probabilistic Trust and Reputation Model for Supply Chain Management.

Paper Link】 【Pages】:

【Authors】: Yasaman Haghpanah

【Abstract】: My thesis contributes to the field of multi-agent systems by proposing a novel trust-based decision model for supply chain management.

【Keywords】:

260. Designing Water Efficient Residential Landscapes with Agent-Based Modeling.

Paper Link】 【Pages】:

【Authors】: Rhonda Hoenigman

【Abstract】: The focus of my research is an agent-based system for optimizing spatial arrangements of plants on a landscape to maximize their growth and minimize their water use. The optimization criteria include a natural phenomenon known as facilitation, which is observed in water-scarce environments when larger shrubs serve as benefactors to smaller annuals by generating conditions that protect them from harsh afternoon sun. In my modeling and optimization system each plant is an agent with growth requirements. A plant agent's fitness at a given location is defined by a fitness function that includes those growth requirements and a penalty term designed to force facilitation. The landscape design is formulated as a combinatorial optimization problem with a discrete set of locations for each plant on a grid, a fixed number of plants, and a fitness function that defines the performance of a plant at a location. To evaluate the effectiveness of this approach, I applied a variety of search strategies, including simulated annealing and a new agent-based approach that mimics how plant communities evolve over time, to different collections of simulated plant types and landscapes and compared the fitness scores and spatial arrangments in the solutions. The fitness scores from the search strategies were comparable. The search strategies produced different spatial distributions of the larger plants, and all designs exhibited facilitation and lower water use.

【Keywords】:

261. Predicting Text Quality for Scientific Articles.

Paper Link】 【Pages】:

【Authors】: Annie Priyadarshini Louis

【Abstract】: My work aims to build a system to automatically predict the writing quality in scientific articles from two genres—academic publications and science journalism. Our goal is to employ these predictions for article recommendation systems and to provide feedback during writing.

【Keywords】:

262. Scaling Up Game Theory: Achievable Set Methods for Efficiently Solving Stochastic Games of Complete and Incomplete Information.

Paper Link】 【Pages】:

【Authors】: Liam MacDermed

【Abstract】: Proposed Thesis: Achievable set methods can efficiently compute useful game theoretic solutions to general multi-agent reinforcement learning problems.

【Keywords】:

Paper Link】 【Pages】:

【Authors】: Nir Pochter

【Abstract】: Search algorithms often suffer from exploring areas which eventually are not part of the shortest path from the start to a goal. Usually it is the purpose of the heuristic function to guide the search algorithm such that it will ignore as much as possible of these areas. We consider other, non-heuristic methods that can be used to prune the search space to make search even faster. We present two algorithms: one for search in graphs that fit in memory, and in which we will need to perform many searches, and another, which improves the search time of planning problems that contain symmetries.

【Keywords】:

264. Learning with Imprecise Classes, Rare Instances, and Complex Relationships.

Paper Link】 【Pages】:

【Authors】: Srinath Ravindran

【Abstract】: In applications including chemoinformatics, bioinfor- matics, information retrieval, text classification, com- puter vision and others, a variety of common issues have been identified involving frequency of occurrence, variation and similarities of instances, and lack of pre- cise class labels. These issues continue to be important hurdles in machine intelligence and my doctoral thesis focuses on developing robust machine learning models that address the same.

【Keywords】:

265. Modeling the Effects of Emotion on Cognition.

Paper Link】 【Pages】:

【Authors】: Marc Spraragen

【Abstract】: Understanding the interaction between emotion and cognitive processes is important for developing architectures for general intelligence, and vital for the fields of human social and behavioral modeling, game intelligence, and human-computer interaction. However, relatively little work in AI has been done on emotion in intelligent architectures, particularly on the effect of emotions on cognitive processes such as inference, planning and learning, despite research showing that emotion is a crucial and often beneficial factor in human decision-making. My work will provide a new emotional-cognitive architecture, focusing on a small set of theories, mechanisms and algorithms for the modeling of a wide array of emotional effects on human cognitive processes. The work and its results will be evaluated against current computational models of cognition and emotion, and validated by results from human cognitive science, neuroscience, and psychology.

【Keywords】:

266. Learning Sensor, Space and Object Geometry.

Paper Link】 【Pages】:

【Authors】: Jeremy Stober

【Abstract】: Robots with many sensors are capable of generating volumes of high-dimensional perceptual data. Making sense of this data and extracting useful knowledge from it is a difficult problem. For robots lacking proper models, trying to understand a stream of uninterpreted data is an especially acute problem. One critical step in linking raw uninterpreted perceptual data to cognition is dimensionality reduction. Current methods for reducing the dimension of data do not meet the demands of a robot situated in the world, and methods that use only perceptual data do not take full advantage of the interactive experience of an embodied robot agent. This work proposes a new scalable, incremental and active approach to dimensionality reduction suitable for extracting geometric knowledge from uninterpreted sensors and effectors. The proposed method uses distinctive state abstractions to organize early sensorimotor experience and sensorimotor embedding to incrementally learn accurate geometric representations based on experience. This approach is applied to the problem of learning the geometry of sensors, space, and objects. The result is evaluated using techniques from statistical shape analysis.

【Keywords】:

267. Incentive-Compatible Trust Mechanisms.

Paper Link】 【Pages】:

【Authors】: Jens Witkowski

【Abstract】: The most prominent way to establish trust in online markets such as eBay are reputation systems that publish buyer feedback about a seller’s past behavior. These systems, however, critically rely on assumptions that are rarely met in realworld marketplaces: first, it is assumed that there are no reporting costs and no benefits from lying so that buyers honestly report their private experiences. Second, it is assumed that every seller is long-lived, i.e. will continue to trade on the marketplace indefinitely and, third, it is assumed that sellers cannot whitewash, i.e. create new accounts once an old one is ran down. In my thesis, I address all of these assumptions and design incentive-compatible trust mechanisms with minimal common knowledge requirements.

【Keywords】:

Student Abstracts and Posters 44

268. Learning Compact Representations of Time-Varying Processes.

Paper Link】 【Pages】:

【Authors】: Philip Bachman ; Doina Precup

【Abstract】: We seek informative representations of the processes underlying time series data. As a first step, we address problems in which these processes can be approximated by linear models that vary smoothly over time. To facilitate estimation of these linear models, we introduce a method of dimension reduction which significantly reduces error when models are estimated locally for each point in time. This improvement is gained by performing dimension reduction implicitly through the model parameters rather than directly in the observation space.

【Keywords】:

269. Assessing Quality in the Web of Linked Sensor Data.

Paper Link】 【Pages】:

【Authors】: Chris Colin Baillie ; Peter Edwards ; Edoardo Pignotti

【Abstract】: Assessing the quality of sensor data available on the Web is essential in order to identify reliable information for decision-making. This paper discusses how provenance of sensor observations and previous quality ratings can influence quality assessment decisions.

【Keywords】:

270. Medical Treatment Conflict Resolving in Answer Set Programming.

Paper Link】 【Pages】:

【Authors】: Forrest Sheng Bao ; Zhizheng Zhang ; Yuanlin Zhang

【Abstract】: Medical treatment decision making is a good application of knowledge representation and reasoning. We are particularly interested in using it to resolve treatment conflicts, a complicated condition when two treatments cannot be given simultaneously to a patient of multiple symptoms. The logic system is required to reason on cases with and without treatment conflicts. Thanks to the nonmonotonicity of Answer Set Programming (ASP), we elegantly automate medical treatment conflict resolving on an example problem and show the importance of nonmonotonicity in medical reasoning.

【Keywords】:

271. Controlling Selection Bias in Causal Inference.

Paper Link】 【Pages】:

【Authors】: Elias Bareinboim ; Judea Pearl

【Abstract】: Selection bias, caused by preferential exclusion of units (or samples) from the data, is a major obstacle to valid causal inferences, for it cannot be removed or even detected by randomized experiments. This paper highlights several graphical and algebraic methods capable of mitigating and sometimes eliminating this bias. These nonparametric methods generalize and improve previously reported results, and identify the type of knowledge that need to be available for reasoning in the presence of selection bias

【Keywords】:

272. Solving 4x5 Dots-And-Boxes.

Paper Link】 【Pages】:

【Authors】: Joseph Kelly Barker ; Richard E. Korf

【Abstract】: Dots-And-Boxes is a well-known and widely-played combinatorial game. While the rules of play are very simple, the state space for even small games is extremely large, and finding the outcome under optimal play is correspondingly hard. In this paper we introduce a Dots-And-Boxes solver which is significantly faster than the current state-of-the-art: over an order-of-magnitude faster on several large problems. We describe our approach, which uses Alpha-Beta search and applies a number of techniques—both problem-specific and general—to reduce the number of duplicate states explored and reduce the search space to a manageable size. Using these techniques, we have determined for the first time that Dots- And-Boxes on a board of 4x5 boxes is a tie given optimal play. This is the largest game solved to date.

【Keywords】:

273. Ad Hoc Teamwork in Variations of the Pursuit Domain.

Paper Link】 【Pages】:

【Authors】: Samuel Barrett ; Peter Stone

【Abstract】: In multiagent team settings, the agents are often given a protocol for coordinating their actions. When such a protocol is not available, agents must engage in ad hoc teamwork to effectively cooperate with one another. A fully general ad hoc team agent needs to be capable of collaborating with a wide range of potential teammates on a varying set of joint tasks. This paper extends previous research in a new direction with the introduction of an efficient method for reasoning about the value of information.  Then, we show how previous theoretical results can aid ad hoc agents in a set of testbed pursuit domains.

【Keywords】:

274. Reconstructing the Stochastic Evolution Diagram of Dynamic Complex Systems.

Paper Link】 【Pages】:

【Authors】: Navid Bazzazzadeh ; Benedikt Brors ; Roland Eils

【Abstract】: The behavior and dynamics of complex systems are in focus of many research fields. The complexity of such systems comes not only from the number of their elements, but also from the unavoidable emergence of new properties of the system, which are not just a simple summation of the properties of its elements. The behavior of complex systems can be fitted with a number of well developed models, which, however, do not incorporate the modularity and the evolution of a system simultaneously. In this work, we propose a generalized model that addresses this issue. Our model is developed within the Random Set Theory’s framework and allows for reconstructing the stochastic evolution diagrams of complex systems.

【Keywords】:

275. Provoking Opponents to Facilitate the Recognition of their Intentions.

Paper Link】 【Pages】:

【Authors】: Francis Bisson ; Froduald Kabanza ; Abder Rezak Benaskeur ; Hengameh Irandoust

【Abstract】: Possessing a sufficient level of situation awareness is essential for effective decision making in dynamic environments. In video games, this includes being aware to some extent of the intentions of the opponents. Such high-level awareness hinges upon inferences over the lower-level situation awareness provided by the game state. Traditional plan recognizers are completely passive processes that leave all the initiative to the observed agent. In a situation where the opponent's intentions are unclear, the observer is forced to wait until further observations of the opponent's actions are made to disambiguate the pending goal hypotheses. With the plan recognizer we propose, in contrast, the observer would take the initiative and provoke the opponent, with the expectation that his reaction will give cues as to what his true intentions actually are.

【Keywords】:

276. Dynamic Batch Mode Active Learning via L1 Regularization.

Paper Link】 【Pages】:

【Authors】: Shayok Chakraborty ; Vineeth Nallure Balasubramanian ; Sethuraman Panchanathan

【Abstract】: We propose a method for dynamic batch mode active learning where the batch size and selection criteria are integrated into a single formulation.

【Keywords】:

277. Heuristic Planning in Adversarial Dynamic Domains.

Paper Link】 【Pages】:

【Authors】: Simon Chamberland ; Froduald Kabanza

【Abstract】: Agents in highly dynamic adversarial domains, such as RTS games, must continually make time-critical decisions to adapt their behaviour to the changing environment. In such a context, the planning agent must consider his opponent's actions as uncontrollable, or at best influenceable. In general nondeterministic domains where there is no clear turn-taking protocol, most heuristic search methods to date do not explicitly reason about the opponent's actions when guiding the state space exploration towards goal or high-reward states. In contrast, we are investigating a domain-independent heuristic planning approach which reasons about the dynamics and uncontrollability of the opponent's behaviours in order to provide better guidance to the search process of the planner. Our planner takes as input the opponent's behaviours recognized by a plan recognition module and uses them to identify opponent's actions that lead to low-utility projected states. We believe such explicit heuristic reasoning about the potential behaviours of the opponent is crucial when planning in adversarial domains, yet is missing in today's planning approaches.

【Keywords】:

Paper Link】 【Pages】:

【Authors】: Hung-Che Chen ; Jyh-Da Wei

【Abstract】: A major difficulty in a search-based problem-solving process is the task of searching the potentially huge search space resulting from the exponential growth of states. State explosion rapidly occupies memory and increases computation time. Although various heuristic search algorithms have been developed to solve problems in a reasonable time, there is no efficient method to construct heuristic functions. In this work, we propose a method by which a neural network can be iteratively trained to form an efficient heuristic function. An adaptive heuristic search procedure is involved in the training iterations. This procedure reduces the evaluation values of the states that are involved in the currently known best solution paths. By doing so, the promising states are continuously moved forward. The adapted heuristic values are fed back to neural networks; thus, a well-trained network function can find the near-best solutions quickly. To demonstrate this method, we solved the fifteen-puzzle problem. Experimental results showed that the solutions obtained by our method were very close to the shortest path, and both the number of explored nodes and the search time were significantly reduced.

【Keywords】:

279. Can Collective Sentiment Expressed on Twitter Predict Political Elections?

Paper Link】 【Pages】:

【Authors】: Jessica Elan Chung ; Eni Mustafaraj

【Abstract】: Research examining the predictive power of social media (especially Twitter) displays conflicting results, particularly in the domain of political elections. This paper applies methods used in studies that have shown a direct correlation between volume/sentiment of Twitter chatter and future electoral results in a new dataset about political elections. We show that these methods display a series of shortcomings that make them inadequate for determining whether social media messages can predict the outcome of elections.

【Keywords】:

280. Conflict-Driven Constraint Answer Set Solving with Lazy Nogood Generation.

Paper Link】 【Pages】:

【Authors】: Christian Drescher ; Toby Walsh

【Abstract】: We present a new approach to enhancing answer set programming (ASP) with constraint programming (CP) techniques based on conflict-driven learning and lazy nogood generation.

【Keywords】:

281. Probabilistic Plan Graph Heuristic for Probabilistic Planning.

Paper Link】 【Pages】:

【Authors】: Yolanda E.-Martín ; María Dolores Rodríguez-Moreno ; David E. Smith

【Abstract】: This work focuses on developing domain-independent heuristics for probabilistic planning problems characterized by full observability and non-deterministic effects of actions that are expressed by probability distributions. The approach is to first search for a high probability deterministic plan using a classical planner. A novel probabilistic plan graph heuristic is used to guide the search towards high probability plans. The resulting plans can be used in a system that handles unexpected outcomes by runtime replanning. The plans can also be incrementally augmented with contingency branches for the most critical action outcomes. This abstract will describe the steps that we have taken in completing the above work and the obtained results.

【Keywords】:

282. Optimal Subset Selection for Active Learning.

Paper Link】 【Pages】:

【Authors】: Yifan Fu ; Xingquan Zhu

【Abstract】: Active learning traditionally relies on instance based utility measures to rank and select instances for labeling, which may result in labeling redundancy. To address this issue, we explore instance utility from two dimensions: individual uncertainty and instance disparity, using a correlation matrix. The active learning is transformed to a semi-definite programming problem to select an optimal subset with maximum utility value. Experiments demonstrate the algorithm performance in comparison with baseline approaches.

【Keywords】:

283. Efficient Issue-Grouping Approach for Multi-Issues Negotiation between Exaggerator Agents.

Paper Link】 【Pages】:

【Authors】: Katsuhide Fujita ; Takayuki Ito ; Mark Klein

【Abstract】: Most real-world negotiation involves multiple interdependent issues, which makes an agent's utility functions complex. Traditional negotiation mechanisms, which were designed for linear utilities, do not fare well in nonlinear contexts. One of the main challenges in developing effective nonlinear negotiation protocols is scalability; it can be extremely difficult to find high-quality solutions when there are many issues, due to computational intractability. One reasonable approach to reducing computational cost, while maintaining good quality outcomes, is to decompose the contract space into several largely independent sub-spaces. In this paper, we propose a method for decomposing a contract space into sub-spaces based on the agent's utility functions. A mediator finds sub-contracts in each sub-space based on votes from the agents, and combines the sub-contracts to produce the final agreement. We demonstrate, experimentally, that our protocol allows high-optimality outcomes with greater scalability than previous efforts. We also address incentive compatibility issues. Any voting scheme introduces the potential for strategic non-truthful voting by the agents, and our method is no exception. For example, one of the agents may always vote truthfully, while the other exaggerates so that its votes are always "strong." It has been shown that this biases the negotiation outcomes to favor the exaggerator, at the cost of reduced social welfare. We employ the limitation of strong votes to the method of decomposing the contract space into several largely independent sub-spaces. We investigate whether and how this approach can be applied to the method of decomposing a contract space.

【Keywords】:

284. Hybrid Tractable Classes of Binary Quantified Constraint Satisfaction Problems.

Paper Link】 【Pages】:

【Authors】: Jian Gao ; Minghao Yin ; Junping Zhou

【Abstract】: In this paper, we investigate the hybrid tractability of binary Quantified Constraint Satisfaction Problems (QCSPs). First, a basic tractable class of binary QCSPs is identified by using the broken-triangle property. In this class, the variable ordering for the broken-triangle property must be same as that in the prefix of the QCSP. Second, we break this restriction to allow that existentially quantified variables can be shifted within or out of their blocks, and thus identify some novel tractable classes by introducing the broken-angle property. Finally, we identify a more generalized tractable class, i.e., the min-of-max extendable class for QCSPs.

【Keywords】:

285. Role-Based Ad Hoc Teamwork.

Paper Link】 【Pages】:

【Authors】: Katie Long Genter ; Noa Agmon ; Peter Stone

【Abstract】: An ad hoc team setting is one in which teammates must work together to obtain a common goal, but without any prior agreement regarding how to work together. In this abstract we present a role-based approach for ad hoc teamwork, in which each teammate is inferred to be following a specialized role that accomplishes a specific task or exhibits a particular behavior. In such cases, the role an ad hoc agent should select depends both on its own capabilities and on the roles currently selected by the other team members. We present methods for evaluating the influence of the ad hoc agent's role selection on the team's utility and we examine empirically how to select the best suited method for role assignment in a complex environment. Finally, we show that an appropriate assignment method can be determined from a limited amount of data and used successfully in similar new tasks that the team has not encountered before.

【Keywords】:

286. Using Conditional Random Fields to Exploit Token Structure and Labels for Accurate Semantic Annotation.

Paper Link】 【Pages】:

【Authors】: Aman Goel ; Craig A. Knoblock ; Kristina Lerman

【Abstract】: Automatic semantic annotation of structured data enables unsupervised integration of data from heterogeneous sources but is difficult to perform accurately due to the presence of many numeric fields and proper-noun fields that do not allow reference-based approaches and the absence of natural language text that prevents the use of language-based approaches. In addition, several of these semantic types have multiple heterogeneous representations, while sharing syntactic structure with other types. In this work, we propose a new approach to use conditional random fields (CRFs) to perform semantic annotation of structured data that takes advantage of the structure and labels of the tokens for higher accuracy of field labeling, while still allowing the use of exact inference techniques. We compare our approach with a linear-CRF based model that only labels fields and also with a regular-expression based approach.

【Keywords】:

287. Large Scale Diagnosis Using Associations between System Outputs and Components.

Paper Link】 【Pages】:

【Authors】: Ting Guo ; Zhanshan Li ; Ruizhi Guo ; Xingquan Zhu

【Abstract】: Model-based diagnosis (MBD) uses an abstraction of system to diagnose possible faulty functions of an underlying system. To improve the solution efficiency for multi-fault diagnosis problems, especially for large scale systems, this paper proposes a method to induce reasonable diagnosis solutions, under coarse diagnosis, by using the relationships between system outputs and components. Compared to existing diagnosis methods, the proposed framework only needs to consider associations between outputs and components by using an assumption-based truth maintenance system (ATMS) [de Kleer 1986] to obtain correlation components for every output node. As a result, our method significantly reduces the number of variables required for model diagnosis, which makes it suitable for large scale circuit systems.

【Keywords】:

288. On the Discovery and Utility of Precedence Constraints in Temporal Planning.

Paper Link】 【Pages】:

【Authors】: Yanmei Hu ; Minghao Yin ; Dunbo Cai

【Abstract】: We extend the precedence constraints contexts heuristic (hpcc) to a temporal and numeric setting, and propose rules to account precedence constraints among comparison variables and logical variables. Experimental results on benchmark domains show that our extension has the potential to lead to better plan quality than that with the heuristic proposed by Eyerich and others.

【Keywords】:

289. Exact Phase Transitions and Approximate Algorithm of #CSP.

Paper Link】 【Pages】:

【Authors】: Ping Huang ; Minghao Yin ; Ke Xu

【Abstract】: The study of phase transition phenomenon of NP complete problems plays an important role in understanding the nature of hard problems. In this paper, we follow this line of research by considering the problem of counting solutions of Constraint Satisfaction Problems (#CSP). We consider the random model, i.e. RB model. We prove that phase transition of #CSP does exist as the number of variables approaches infinity and the critical values where phase transitions occur are precisely located. Preliminary experimental results also show that the critical point coincides with the theoretical derivation. Moreover, we propose an approximate algorithm to estimate the expectation value of the solutions number of a given CSP instance of RB model.

【Keywords】:

290. Extending the Applications of Recent Real-Time Heuristic Search.

Paper Link】 【Pages】:

【Authors】: Daniel Andrew Huntley ; Vadim Bulitko

【Abstract】: Real-time heuristic search algorithms that precompute search space-specific databases have demonstrated exceptional performance in video-game pathfinding. We discuss the first steps towards extending these algorithms to other search spaces that also benefit from the real-time property. We present our initial progress in characterizing the performance of current algorithms based on the features of a search space, and discuss future directions of this research.

【Keywords】:

291. Multiple-Instance Learning: Multiple Feature Selection on Instance Representation.

Paper Link】 【Pages】:

【Authors】: I-Hong Jhuo ; D. T. Lee

【Abstract】: In multiple-Instance Learning (MIL), training class labels are attached to sets of bags composed of unlabeled instances, and the goal is to deal with classification of bags. Most previous MIL algorithms, which tackle classification problems, consider each instance as a represented feature. Although the algorithms work well in some prediction problems, considering diverse features to represent an instance may provide more significant information for learning task. Moreover, since each instance may be mapped into diverse feature spaces, encountering a large number of irrelevant or redundant features is inevitable. In this paper, we propose a method to select relevant instances and concurrently consider multiple features for each instance, which is termed as MIL-MFS. MIL-MFS is based on multiple kernel learning (MKL), and it iteratively selects the fusing multiple features for classifier training. Experimental results show that the MIL-MFS combined with multiple kernel learning can significantly improve the classification performance.

【Keywords】:

292. An Event-Based Framework for Process Inference.

Paper Link】 【Pages】:

【Authors】: Michael Joya

【Abstract】: We focus on a class of models used for representing the dynamics between a discrete set of probabilistic events in a continuous-time setting. The proposed framework offers tractable learning and inference procedures and provides compact state representations for processes which exhibit variable delays between events. The approach is applied to a heart sound labeling task that exhibits long-range dependencies on previous events, and in which explicit modeling of the rhythm timings is justifiable by cardiological principles.

【Keywords】:

293. Toward Learning to Solve Insertion Tasks: A Developmental Approach Using Exploratory Behaviors and Proprioception.

Paper Link】 【Pages】:

【Authors】: Philip Koonce ; Vasha Dutell ; Jose Farrington ; Vladimir Sukhoy ; Alexander Stoytchev

【Abstract】: This paper describes an approach to solving insertion tasks by a robot that uses exploratory behaviors and proprioceptive feedback. The approach was inspired by the developmental progression of insertion abilities in both chimpanzees and humans (Hayashi et al. 2006). Before mastering insertions, the infants of the two species undergo a stage where they only press objects against other objects without releasing them. Our goal was to emulate this developmental stage on a robot to see if it may lead to simpler representations for insertion tasks. Experiments were performed using a shapesorter puzzle with three different blocks and holes.

【Keywords】:

294. Time Complexity of Iterative-Deepening A*: The Informativeness Pathology (Abstract).

Paper Link】 【Pages】:

【Authors】: Levi Lelis ; Sandra Zilles ; Robert C. Holte

【Abstract】: Korf, Reid, and Edelkamp launched a line of research aimed at predicting how many nodes IDA* will expand with a given depth bound. This paper advances this line of research in three ways. First, we identify a source of prediction error that has hitherto been overlooked. We call it the "discretization effect." Second, we disprove the intuitively appealing idea that a "more informed" prediction system cannot make worse predictions than a ``less informed'' one. More informed systems are more susceptible to the discretization effect, and in our experiments the more informed system makes poorer predictions. Our third contribution is a method, called "Epsilon-truncation," which makes a prediction system less informed, in a carefully chosen way, so as to improve its predictions by reducing the discretization effect. In our experiments Epsilon-truncation improved predictions substantially.

【Keywords】:

295. An Empirical Study of Bagging Predictors for Different Learning Algorithms.

Paper Link】 【Pages】:

【Authors】: Guohua Liang ; Xingquan Zhu ; Chengqi Zhang

【Abstract】: Bagging is a simple yet effective design which combines multiple single learners to form an ensemble for prediction. Despite its popular usage in many real-world applications, existing research is mainly concerned with studying unstable learners as the key to ensure the performance gain of a bagging predictor, with many key factors remaining unclear. For example, it is not clear when a bagging predictor can outperform a single learner and what is the expected performance gain when different learning algorithms were used to form a bagging predictor. In this paper, we carry out comprehensive empirical studies to evaluate bagging predictors by using 12 different learning algorithms and 48 benchmark data-sets. Our analysis uses robustness and stability decompositions to characterize different learning algorithms, through which we rank all learning algorithms and comparatively study their bagging predictors to draw conclusions. Our studies assert that both stability and robustness are key requirements to ensure the high performance for building a bagging predictor. In addition, our studies demonstrated that bagging is statistically superior to most single base learners, except for KNN and Naïve Bayes (NB). Multi-layer perception (MLP), Naïve Bayes Trees (NBTree), and PART are the learning algorithms with the best bagging performance.

【Keywords】:

296. An Efficient and Complete Approach for Cooperative Path-Finding.

Paper Link】 【Pages】:

【Authors】: Ryan Luna ; Kostas E. Bekris

【Abstract】: Cooperative path-finding can be abstracted as computing non-colliding paths for multiple agents between their start and goal locations on a graph. This work proposes a fast algorithm that can provide completeness guarantees for a general class of problems without any assumptions about the graph's topology. Specifically, the approach can address any solvable instance where there are at most n-2 agents in a graph of size n. The algorithm employs two primitives: a "push" operation where agents move towards their goals up to the point that no progress can be made, and a "swap" operation that allows two agents to swap positions without altering the configuration of other agents. Simulated experiments are provided on hard instances of cooperative path-finding, including comparisons against alternative methods. The results are favorable for the proposed algorithm and show that the technique scales to problems that require high levels of coordination, involving hundreds of agents.

【Keywords】:

297. Generating Explanations for Complex Biomedical Queries.

Paper Link】 【Pages】:

【Authors】: Umut Öztok ; Esra Erdem

【Abstract】: We present a computational method to generate explanations to answers of complex queries over biomedical ontologies and databases, using the high-level representation and efficient automated reasoners of Answer Set Programming. We show the applicability of our approach with some queries related to drug discovery over PHARMGKB, DRUGBANK, BIOGRID, CTD and SIDER.

【Keywords】:

298. An Intelligent System for Prolonging Independent Living of Elderly.

Paper Link】 【Pages】:

【Authors】: Bogdan Pogorelc

【Abstract】: The number of elderly people is constantly increasing in the developed countries. Elderly tend to lead an isolated life away from their offspring; however, they may fear being unable to obtain help if they are injured or ill. During the last decades, this fear has generated research attempts to find assistive technologies for making living of elderly people at homes easier and independent, as is the aim of this research work. Research study proposes a generalized approach to an intelligent and ubiquitous care system to recognize a few of the most common and important health problems of the elderly, which can be detected by analyzing their movement. In the event that the system was to recognize a health problem, it would automatically notify a physician with an included explanation of the automatic diagnosis. It is two-step approach; in the first step it classifies person's activities into five activities: fall, unconscious fall, walking, standing/sitting, lying down/lying. In the second step, it classifies walking patterns into five different health states; one healthy and four unhealthy: hemiplegia (usually the result of stroke), Parkinson’s disease, leg pain and back pain. Moreover, since elderly having these health problems are less stable and more prone to falls, recognizing them leads not only to detection but indirectly also to prevention of falls of elderly people. In the initial approach movement of the user is captured with the motion capture system, which consists of the tags attached to the body, whose coordinates are acquired by the sensors situated in the apartment. In the current approach wearable inertial sensors are used, allowing monitoring inside or outside of the buildings. Output time-series of coordinates are modeled with the proposed data mining approach to recognize the specific health problem.

【Keywords】:

Paper Link】 【Pages】:

【Authors】: Baojun Qiu ; Qi He ; John Yen

【Abstract】: Link prediction is one of central tasks in the study of social network evolution and has many applications. In this paper, we use time series to describe node behavior, extract temporal features from the time series to characterize behavior evolution of nodes, and use the temporal features for link prediction. Our experimental results on several real datasets suggest that including the temporal features developed in the paper significantly improve link prediction performance.

【Keywords】:

300. Web Personalization and Cohort Information Services for Natural Resource Managers.

Paper Link】 【Pages】:

【Authors】: Crystal Redman

【Abstract】: We aim to address yet unanswered questions in web personalization by providing an integrated view of documents found through different tools, considering user confidentiality, emphasizing the benefits of client-side web personalization, and highlighting the power of social webpage recommendation when traditional information retrieval and collaborative filtering methods are used in conjunction. These principles define a framework which leverages groupings of users with overlapping interests called user cohorts.

【Keywords】:

301. Using Partitions and Superstrings for Lossless Compression of Pattern Databases.

Paper Link】 【Pages】:

【Authors】: Ethan L. Schreiber ; Richard E. Korf

【Abstract】: We present an algorithm for compressing pattern databases (PDBs) and a method for fast random access of these com-pressed PDBs. We demonstrate the effectiveness of our technique by compressing two 6-tile sliding-tile PDBs by a factor of 12 and a 7-tile sliding-tile PDB by a factor of 24.

【Keywords】:

302. Convergence Properties of (μ + λ) Evolutionary Algorithms.

Paper Link】 【Pages】:

【Authors】: Aram Ter-Sarkisov ; Stephen R. Marsland

【Abstract】: We present a number of convergence properties of population-based Evolutionary Algorithms (EAs) on a set of test functions. Focus is on EA using k-Bit-Swap (kBS) operator. We compare our findings to past research.  

【Keywords】:

303. On the Effectiveness of Belief State Representation in Contingent Planning.

Paper Link】 【Pages】:

【Authors】: Son Thanh To ; Tran Cao Son ; Enrico Pontelli

【Abstract】: This work proposes new approaches to contingent planning using alternative belief state representations extended from those in conformant planning and a new AND/OR forward search algorithm, called PrAO, for contingent solutions. Each representation was implemented in a new contingent planner. The important role of belief state representation has been confirmed by the fact that our planners all outperform other stateof- the-art planners on most benchmarks and the comparison of their performances varies across all the benchmarks even using the same search algorithm PrAO and same unsophisticated heuristic scheme. The work identifies the properties of each representation method that affect the performance.

【Keywords】:

304. A Bayesian Reinforcement Learning framework Using Relevant Vector Machines.

Paper Link】 【Pages】:

【Authors】: Nikolaos Tziortziotis ; Konstantinos Blekas

【Abstract】: In this work we present an advanced Bayesian formulation to the task of control learning that employs the Relevance Vector Machines (RVM) generative model for value function evaluation. The key aspect of the proposed method is the design of the discount return as a generalized linear model that constitutes a well-known probabilistic approach. This allows to augment the model with advantageous sparse priors provided by the RVM's regression framework. We have also taken into account the significant issue of selecting the proper parameters of the kernel design matrix. Experiments have shown that our method produces improved performance in both simulated and real test environments.

【Keywords】:

305. A Framework for Integration of Logical and Probabilistic Knowledge.

Paper Link】 【Pages】:

【Authors】: Jingsong Wang ; Marco Valtorta

【Abstract】: Integrating the expressive power of first-order logic with the ability of probabilistic reasoning of Bayesian networks has attracted the interest of many researchers for decades. We present an approach to integration that translates logical knowledge into Bayesian networks and uses Bayesian network composition to build a uniform representation that supports both logical and probabilistic reasoning. In particular, we propose a new way of translation of logical knowledge, relation search. Through the use of the proposed framework, without learning new languages or tools, modelers are allowed to 1) specify special knowledge using the most suitable languages, while reasoning in a uniform engine; 2) make use of pre-existing logical knowledge bases for probabilistic reasoning (to complete the model or minimize potential inconsistencies).

【Keywords】:

306. Solution Quality Improvements for Massively Multi-Agent Pathfinding.

Paper Link】 【Pages】:

【Authors】: Ko-Hsin Cindy Wang ; Adi Botea ; Philip Kilby

【Abstract】: MAPP has been previously shown as a state-of-the-art multi-agent path planning algorithm on criteria including scalability and success ratio (i.e., percentage of solved units) on realistic game maps. MAPP further provides a formal characterization of problems it can solve, and low-polynomial upper bounds on the resources required. However, until now, MAPP's solution quality had not been extensively analyzed. In this work we empirically analyze the quality of MAPP's solutions, using multiple quality criteria such as the total travel distance, the makespan and the sum of actions (including move and wait actions). We also introduce enhancements that improve MAPP's solution quality significantly. For example, the sum of actions is cut to half on average. The improved MAPP is competitive in terms of solution quality with FAR and WHCA*, two successful algorithms from the literature, and maintains its advantages on different performance criteria, such as scalability, success ratio, and ability to tell apriori if it will succeed in the instance at hand. As optimal algorithms have limited scalability, evaluating the quality of the solutions provided by suboptimal algorithms is another important topic. Using lower bounds of optimal values, we show that MAPP's solutions have a reasonable quality. For example, MAPP's total travel distance is on average 19% longer than a lower bound on the optimal value.

【Keywords】:

307. Online Updating the Generalized Inverse of Centered Matrices.

Paper Link】 【Pages】:

【Authors】: Qing Wang ; Liang Zhang

【Abstract】: In this paper, we present the exact online updating formulae for the generalized inverse of centered matrices. The computational cost is O ( mn ) for matrices of size m × n . Experimental results validate the proposed method’s accuracy and efficiency.  

【Keywords】:

308. Modeling Opponent Actions for Table-Tennis Playing Robot.

Paper Link】 【Pages】:

【Authors】: Zhikun Wang ; Abdeslam Boularias ; Katharina Mülling ; Jan Peters

【Abstract】: Opponent modeling is a critical mechanism in repeated games. It allows a player to adapt its strategy in order to better respond to the presumed preferences of its opponents. We introduce a modeling technique that adaptively balances safety and exploitability. The opponent's strategy is modeled with a set of possible strategies that contains the actual one with high probability. The algorithm is safe as the expected payoff is above the minimax payoff with high probability, and can exploit the opponent's preferences when sufficient observations are obtained. We apply the algorithm to a robot table-tennis setting where the robot player learns to prepare to return a served ball. By modeling the human players, the robot chooses a forehand, backhand or middle preparation pose before they serve. The learned strategies can exploit the opponent's preferences, leading to a higher rate of successful returns.

【Keywords】:

309. Adaptive Neighborhood Inverse Consistency as Lookahead for Non-Binary CSPs.

Paper Link】 【Pages】:

【Authors】: Robert J. Woodward ; Shant Karakashian ; Berthe Y. Choueiry ; Christian Bessiere

【Abstract】: Freuder and Elfe (1996) introduced Neighborhood Inverse Consistency (NIC) for binary CSPs. In this paper, we introduce RNIC, the extension of NIC to non-binary CSPs, and describe a practical algorithm for enforcing it. We propose an adaptive strategy to weaken or strengthen this property based on the connectivity of the network. We demonstrate the effectiveness of RNIC as a full lookahead strategy during search for solving difficult benchmark problems.

【Keywords】:

Paper Link】 【Pages】:

【Authors】: Fan Xie ; Hootan Nakhost ; Martin Müller

【Abstract】: Much recent work in satisficing planning has aimed at striking a balance between coverage - solving as many problems as possible - and plan quality. Current planners achieve near perfect coverage on the latest IPC benchmarks. It is therefore natural to investigate their scaling behavior on more difficult instances. Among state of the art planners, LAMA (Richter, Helmert, and Westphal 2008) is able to generate high quality plans, but its coverage drops off rapidly with increasing prob- lem complexity. The Arvand planner (Nakhost and Müller 2009) scales to much harder instances but generates lower quality plans. This paper introduces a new algorithm, Monte Carlo Random Walk-based Local Tree Search (MRW-LTS), which uses random walks to selectively build local search trees. Experiments demonstrate that MRW-LTS combines a scaling behavior that is better than LAMA’s with a plan quality that is better than Arvand’s.

【Keywords】:

311. Discovering Latent Strategies.

Paper Link】 【Pages】:

【Authors】: Xiaoxi Xu

【Abstract】: Strategy mining is a new area of research about discovering strategies in decision-making. In this paper, we formulate the strategy-mining problem as a clustering problem, called the latent-strategy problem. In a latent-strategy problem, a corpus of data instances is given, each of which is represented by a set of features and a decision label. The inherent dependency of the decision label on the features is governed by a latent strategy. The objective is to find clusters, each of which contains data instances governed by the same strategy. Existing clustering algorithms are inappropriate to cluster dependency because they either assume feature independency (e.g., K-means) or only consider the co-occurrence of features without explicitly modeling the special dependency of the decision label on other features (e.g., Latent Dirichlet Allocation (LDA)). In this paper, we present a baseline unsupervised learning algorithm for dependency clustering. Our model-based clustering algorithm iterates between an assignment step and a minimization step to learn a mixture of decision tree models that represent latent strategies. Similar to the Expectation Maximization algorithm, our algorithm is grounded in the statistical learning theory. Different from other clustering algorithms, our algorithm is irrelevant-feature resistant and its learned clusters (modeled by decision trees) are strongly interpretable and predictive. We systematically evaluate our algorithm using a common law dataset comprised of actual cases. Experimental results show that our algorithm significantly outperforms K-means and LDA on clustering dependency.

【Keywords】:

Robotics Program 8

312. Learning Tasks and Skills Together From a Human Teacher.

Paper Link】 【Pages】:

【Authors】: Baris Akgun ; Kaushik Subramanian ; Jaeeun Shim ; Andrea Lockerd Thomaz

【Abstract】: We are interested in developing Learning from Demonstration (LfD) systems that are tailored to be used by everyday people. We highlight and tackle the issues of skill learning, task learning and interaction in the context of LfD As part of the AAAI 2011 LfD Challenge, we will demonstrate some of our most recent Socially Guided-Machine Learning work, in which the PR2 robot learns both low-level skills and high-level tasks through an ongoing social dialog with a human partner

【Keywords】:

313. Can Quadrotors Succeed as an Educational Platform?

Paper Link】 【Pages】:

【Authors】: Zachary Dodds

【Abstract】: The flexibility and controllability of quadrotor helicopters have made them a recent focus of interest among robotics and AI research groups. At the same time, their popularity has led to a wide range of commercially available platforms, some at prices accessible for undergraduate educational use. This project evaluates the ARDrone quadrotor helicopter as a basis for use in undergraduate classes such as robotics, computer vision, or embodied AI. We have encountered both successes and frustrations in using the ARDrone to date. Looking forward, the quadrotor’s capabilities do seem a promising basis for future curricular offerings.

【Keywords】:

314. Playing Chess with a Human-Scale Mobile Manipulator.

Paper Link】 【Pages】:

【Authors】: Michael Ferguson ; Kimberly A. Gero ; Joao Salles ; James Weis

【Abstract】: This paper describes our efforts preparing a mobile manipulator for the 2011 AAAI Small Scale Manipulation Challenge. We describe our approach to building a low-cost mobile manipulator for human-scale environments using ROS and off-the-shelf sensory and servos.

【Keywords】:

315. A Robotics Environment for Software Engineering Courses.

Paper Link】 【Pages】:

【Authors】: Stephan Goebel ; Ruben Jubeh ; Simon-Lennert Raesch

【Abstract】: The initial idea of using Lego Mindstorms Robots for student courses had soon to be expanded to a simulation environment as the user base in students grew larger and the need for parallel development and testing arose. An easy to use and easy to set up means of providing positioning data led to the creation of an indoor positioning system so that new users can adapt quickly and successfully, as sensors on the actual robots are difficult to configure and hard to interpret in an environmental context. A global positioning system shared among robots can make local sensors obsolete and still deliver more precise information than currently available sensors, also providing the base necessary for the robots to effectively work on shared tasks as a group. Further more, a simulator for robots programmed with Fujaba and Java which was developed along the way can be used by many developers simultaneously and lets them evaluate their code in a simple way, while close to real-world results.

【Keywords】:

316. Lego Plays Chess: A Low-Cost, Low-Complexity Approach to Intelligent Robotics.

Paper Link】 【Pages】:

【Authors】: Michael Lanighan ; Jerod Sikorskyj ; Debra T. Burhans ; Robert Selkowitz

【Abstract】: The design and implementation of a robotic chess agent is described. Shallow Blue, a competitor in the AAAI 2011 Small Scale Manipulation Challenge, is constructed with low-cost components including Lego NXT bricks and is programmed using Java and Lejos.

【Keywords】:

317. Learning from Demonstration in Spatial Exploration.

Paper Link】 【Pages】:

【Authors】: Juan Pablo Munoz ; Arif Tuna Ozgelen ; Elizabeth Sklar

【Abstract】: We present the initial stage of our research on Learning from Demonstration algorithms. We have implemented an algorithm based on Confident Execution, one of the components of the Confidence-Based Autonomy algorithm developed by Chernova and Veloso. Our preliminary experiments were conducted first in simulation and then using a Sony AIBO ERS-7 robot. So far, our robot has been able to learn crude navigation strategies, despite limited trials. We are currently working on improving our implementation by including additional features that describe more broadly the state of the agent. Our long term goal is to incorporate Learning from Demonstration techniques in our HRTeam (human/multi-robot) framework.

【Keywords】:

318. Approaches to Multi-Robot Exploration and Localization.

Paper Link】 【Pages】:

【Authors】: Arif Tuna Ozgelen ; Michael Costantino ; Adiba Ishak ; Moses Kingston ; Diquan Moore ; Samuel Sanchez ; Juan Pablo Munoz ; Simon Parsons ; Elizabeth Sklar

【Abstract】: We present approaches to several fundamental tasks in multi-robot team-based exploration and localization, based on student projects developed in the past year.

【Keywords】:

319. Hekateros: A Desktop 5 Degree-of-Freedom Robot Arm for the Small-Scale Manipulation Robot Chess Challenge.

Paper Link】 【Pages】:

【Authors】: Kim Wheeler ; Robin Knight ; Collin Horvat ; Daniel Packard ; Casey Kuhns ; Brent Wilkins ; Robert Shiely

【Abstract】: RoadNarrows has entered the “desktop-size” 5 degree of freedom robot arm, Hekateros, in the Small-Scale Manipulator Challenge, at the 2011 AAAI conference. Hekateros is utilizing a variety of sensors and software to successfully perceive and manipulate the chess gam

【Keywords】: