Posts Tagged ‘natural language’

Domain Ontology Construction from Biomedical Text

NLM’s Unified Medical Language System (UMLS) is a very large ontology of biomedical and health data. In order to be used effectively for knowledge processing, it needs to be customized to a specific domain. In this paper, we present techniques to automatically discover domain-specific concepts, discover relationships between these concepts, build a context map from these relationships, link these domain concepts with the best-matching concept identifiers in UMLS using our context map and UMLS concept trees, and finally assign categories to the discovered relationships. This specific domain ontology of terms and relationships using evidential information can serve as a basis for applications in analysis, reasoning and discovery of new relationships. We have automatically built an ontology for the Nuclear Cardiology domain as a testbed for our techniques.

Read the paper:

Domain Ontology Construction from Biomedical Text

by Saurav Sahay, Baoli Li, Ernie Garcia, Eugene Agichtein, Ashwin Ram

International Conference on Artificial Intelligence (ICAI-07), Las Vegas, NV, June 2007
www.cc.gatech.edu/faculty/ashwin/papers/er-07-10.pdf

Emotionally Driven Natural Language Generation for Personality Rich Characters in Interactive Games

Natural Language Generation for personality rich characters represents one of the important directions for believable agents research. The typical approach to interactive NLG is to hand-author textual responses to different situations. In this paper we address NLG for interactive games. Specifically, we present a novel template-based system that provides two distinct advantages over existing systems. First, our system not only works for dialogue, but enables a character’s personality and emotional state to influence the feel of the utterance. Second, our templates are resuable across characters, thus decreasing the burden on the game author. We briefly describe our system and present results of a preliminary evaluation study.

Read the paper:

Emotionally Driven Natural Language Generation for Personality Rich Characters in Interactive Games

by Christina Strong, Kinshuk Mishra, Manish Mehta, Alistair Jones, Ashwin Ram

Third Conference on Artificial Intelligence for Interactive Digital Entertainment (AIIDE-07), Stanford, CA, June 2007
www.cc.gatech.edu/faculty/ashwin/papers/er-07-09.pdf

Evaluating Player Modeling for a Drama Manager Based Interactive Fiction

A growing research community is working towards employing drama management components in story-based games that guide the story towards specific narrative arcs depending on a particular player’s playing patterns. Intuitively, player modeling should be a key component for Drama Manager (DM) based approaches to succeed with human players.

In this paper, we report a particular implementation of the DM component connected to an interactive story game, Anchorhead, while specifically focusing on the player modeling component. We analyze results from our evaluation study and show that similarity in the trace of DM decisions in previous games can be used to predict interestingness of game events for the current player. Results from our current analysis indicate that the average time spent in performing player actions provides a strong distinction between players with varying degrees of gaming experience, thereby helping the DM to adapt its strategy based on this information.

Read the paper:

Evaluating Player Modeling for a Drama Manager Based Interactive Fiction

by Manu Sharma, Manish Mehta, Santi Ontañón, Ashwin Ram

Third Conference on Artificial Intelligence for Interactive Digital Entertainment (AIIDE-07), Workshop on Player Satisfaction, Stanford, CA, June 2007
www.cc.gatech.edu/faculty/ashwin/papers/er-07-08.pdf

Detecting Medical Rule Sentences with Semi-Automatically Derived Patterns: A Pilot Study

We propose a semi-supervised method to extract rule sentences from medical abstracts. Medical rules are sentences that give interesting and non-trivial relationship between medical entities. Mining such medical rules is important since the rules thus extracted can be used as inputs to an expert system or in many more other ways. The technique we suggest is based on paraphrasing a set of seed sentences and populating a pattern dictionary of paraphrases of rules. We match the patterns against the new abstract and rank the sentences.

Read the paper:

Detecting Medical Rule Sentences with Semi-Automatically Derived Patterns: A Pilot Study

by Shreekanth Karvaje, Bharat Ravisekar, Baoli Li, Ernie Garcia, Ashwin Ram

International Symposium on Bioinformatics Research and Applications ( ISBRA-07), Atlanta, GA, May 2007
www.cc.gatech.edu/faculty/ashwin/papers/er-07-07.pdf

Text Mining Biomedical Literature for Discovering Gene-to-Gene Relationships

Partitioning closely related genes into clusters has become an important element of practically all statistical analyses of microarray data. A number of computer algorithms have been developed for this task. Although these algorithms have demonstrated their usefulness for gene clustering, some basic problems remain. This paper describes our work on extracting functional keywords from MEDLINE for a set of genes that are isolated for further study from microarray experiments based on their differential expression patterns. The sharing of functional keywords among genes is used as a basis for clustering in a new approach called BEA-PARTITION. Functional keywords associated with genes were extracted from MEDLINE abstracts. We modified the Bond Energy Algorithm (BEA), which is widely accepted in psychology and database design but is virtually unknown in bioinformatics, to cluster genes by functional keyword associations.

The results showed that BEA-PARTITION and hierarchical clustering algorithm outperformed k-means clustering and self-organizing map by correctly assigning 25 of 26 genes in a test set of four known gene groups. To evaluate the effectiveness of BEA-PARTITION for clustering genes identified by microarray profiles, 44 yeast genes that are differentially expressed during the cell cycle and have been widely studied in the literature were used as a second test set. Using established measures of cluster quality, the results produced by BEA-PARTITION had higher purity, lower entropy, and higher mutual information than those produced by k-means and self-organizing map. Whereas BEA-PARTITION and the hierarchical clustering produced similar quality of clusters, BEA-PARTITION provides clear cluster boundaries compared to the hierarchical clustering. BEA-PARTITION is simple to implement and provides a powerful approach to clustering genes or to any clustering problem where starting matrices are available from experimental observations.

Text Mining Biomedical Literature for Discovering Gene-to-Gene Relationships

by Ying Liu, Sham Navathe, Jorge Civera, Venu Dasigi, Ashwin Ram, Brian Ciliax, Ray Dingledine

IEEE/ACM Transactions on Computational Biology and Bioinformatics,2(4):380-384, Oct-Dec 2005
www.cc.gatech.edu/faculty/ashwin/papers/er-05-01.pdf

Preventing Failures by Mining Maintenance Logs with Case-Based Reasoning

The project integrates work in natural language processing, machine learning, and the semantic web, bringing together these diverse disciplines in a novel way to address a real problem. The objective is to extract and categorize machine components and subsystems and their associated failures using a novel approach that combines text analysis, unsupervised text clustering, and domain models. Through industrial partnerships, this project will demonstrate effectiveness of the proposed approach with actual industry data.

Read the paper:

Preventing Failures by Mining Maintenance Logs with Case-Based Reasoning

by Mark Devaney, Ashwin Ram, Hai Qui, Jay Lee

59th Meeting of the Society for Machinery Failure Prevention Technology (MFPT-59), Virginia Beach, VA, April 2005
www.cc.gatech.edu/faculty/ashwin/papers/er-05-04.pdf

Interactive Case-Based Reasoning for Precise Information Retrieval

The knowledge explosion has continued to outpace technological innovation in search engines and knowledge management systems. It is increasingly difficult to find relevant information, not just on the World Wide Web at large but even in domain- specific medium-sized knowledge bases—online helpdesks, maintenance records, technical repositories, travel databases, e-commerce sites, and many others. Despite advances in search and database technology, the average user still spends inordinate amounts of time looking for specific information needed for a given task.

This paper describes an adaptive system for the precise, rapid retrieval and synthesis of information from medium-sized knowledge bases in response to problem-solving queries from a diverse user population. We advocate a shift in perspective from “search” to “answers. Instead of returning dozens or hundreds of hits to a user, the system should attempt to find answers that may or may not match the query directly but are relevant to the user’s problem or task.

This problem has been largely overlooked as research has tended to concentrate on techniques for broad searches of large databases over the Internet (as exemplified by Google) and structured queries of well-defined databases (as exemplified by SQL). However, the problem discussed in this chapter is sufficiently different from these extremes to both present a novel set of challenges as well as provide a unique opportunity to apply techniques not traditionally found in the information retrieval literature. Specifically, we discuss an innovative combination of techniques‚ case-based reasoning coupled with text analytics‚ to solve the problem in a practical, real-world context.

We are interested in applications in which users must quickly retrieve answers to specific questions or problems from a complex information database with a minimum of effort and interaction. Examples include internal helpdesk support, web-based self-help for consumer products, decision-aiding systems for support personnel, and repositories for specialized documents such as patents, technical documents, or scientific literature. These applications are characterized by the fact that a diverse user population accesses highly focused knowledge bases in order to find precise answers to specific questions or problems. Despite the growing popularity of on-line service and support facilities for internal use by employees and for external use for customers, most such sites rely on traditional search engine technologies and are not very effective in reducing the time, expertise, and complexity required on the user’s part.

Read the paper:

Interactive Case-Based Reasoning for Precise Information Retrieval

by Ashwin Ram, Mark Devaney

In Case-Based Reasoning in Knowledge Discovery and Data Mining, David Aha and Sankar Pal (editors).
www.cc.gatech.edu/faculty/ashwin/papers/er-05-02.pdf

IRIA: The Information Research Intelligent Assistant

The explosion of information in the modern environment demands the ability to collect, organize, manage, and search large amounts of information across a wide variety of real-world applications. The primary tools available for such tasks are large-scale database systems and keyword-based document search techniques. However, such tools are rapidly proving inadequate: traditional database systems do not enable ready access to relevant knowledge, prompting a market of add-ons and existing search techniques are insufficiently precise or selective to support such tasks, leading to consumer exasperation. In the end users are left unsatisfied, confronted with a sea of unorganized and unhelpful data. A new approach is needed.

The Information Research Intelligent Assistant (IRIA) is an integrated information retrieval architecture that addresses this problem. IRIA enables a user or workgroup to build a personalized map of the relevant information available in a database, intranet, or internet, and the ability to find, add, and use information quickly and easily. An IRIA-based intelligent information management system acts as an autonomous assistant to a user working on a task, working unobtrusively in the background to learn both the user’s interests and the resources available to satisfy those interests. This approach enables “reminding engines” which monitor a user’s work to proactively find and recommend useful information as well as “workgroup memories” which learn from a user’s behavior to build a comprehensive knowledge map of a particular area of interest.

In empirical tests, IRIA has demonstrated the ability to monitor a user’s progress on a task (specifically, web search) and proactively find and recommend information relevant to that task based on the context and history of the user’s interactions with the system. IRIA further demonstrated that it could provide collaborative facilities to the workgroup and that it could learn and improve its knowledge map over time.

Read the paper:

IRIA: The Information Research Intelligent Assistant

by Anthony Francis, Mark Devaney, Ashwin Ram

International Conference on Artificial Intelligence (ICAI-00), Las Vegas, Nevada
www.dresan.com/research/publications/icai-2000.html

Context-Sensitive Asynchronous Memory

Retrieving useful answers from large knowledge bases given under-specified questions is an important problem in the construction of general intelligent agents. The core of this problem is how to get the information an agent needs when it doesn’t know how to ask the right question and doesn’t have the time to exhaustively search all available information.

Context-sensitive asynchronous memory is a model of memory retrieval that solves this problem. The context-sensitive asynchronous memory approach exploits feedback from the task and environment to guide and constrain memory search by interleaving memory retrieval and problem solving. To achieve this behavior, a context-sensitive asynchronous memory uses an asynchronous retrieval system to manage a context- sensitive search process operating over a content-addressable knowledge base. Solutions based on this approach provide useful answers to vague questions efficiently, based on information naturally available during the performance of a task.

The core claims of this approach are:
•  Claim 1: An efficient, domain-independent solution to the problem of retrieving useful answers from large knowledge bases given under-specified queries is to interleave memory retrieval with task performance and use feedback from the task or environment to guide the search of memory.
•  Claim 2: Interleaving memory retrieval with and exploiting feedback from task performance can be achieved in a domain-independent way using a context- sensitive, asynchronous memory retrieval process.
•  Claim 3: A rich, reified, grounded semantic network representation enables context-sensitive memory retrieval processes to retrieve useful information in a domain-independent way for a wide variety of tasks.
•  Claim 4: To effectively use a context-sensitive asynchronous memory to retrieve useful answers, a task must be able to work in parallel with a memory process, communicate with it, provide feedback to it, and must possess integration mechanisms to incorporate asynchronous retrievals provided by the memory.

The context-sensitive asynchronous memory approach is applicable to tasks and domains which exhibit the following criteria: problems are difficult to solve, questions are difficult to formulate, a large knowledge base is available yet contains only a small selection of relevant information, and, most importantly, the environment is regular, in that solutions in the knowledge base occur in patterns and relationships similar to those found in situations in which the solutions are likely to be applicable in the future. This approach is domain independent: it is applicable to a wide variety of tasks and problems from simple search applications to complex cognitive agents.

To exploit context-sensitive asynchronous memory, reasoners need certain properties. Experience-based agency is an agent architecture which provides an outline of how to construct complete intelligent agents which use a context-sensitive asynchronous memory to support a reasoning system performing a real task. The experience-based agent architecture combines a context-sensitive asynchronous memory retrieval process with a global store of experience used by all agent processes, a global working memory to provide a uniform way to collect feedback, and a global task controller which orchestrates reasoning and memory. The experience-based agent architecture also provides principles for constructing integration mechanisms that enable reasoning tasks to work with the context-sensitive asynchronous memory.

Furthermore, to help determine when these approaches should be used, this research also contributes theoretical analyses that predict the classes of tasks and situations in which the context-sensitive asynchronous memory and experience-based agent approaches will provide the greatest benefit.

To evaluate the approach, the experience-based agent architecture has been implemented in the Nicole system. Nicole is a large Common Lisp program providing global long-term and working memory stores represented as a rich, reified, grounded semantic network, a context-sensitive asynchronous memory process based on a novel model of context-directed spreading activation, a control system for orchestrating reasoning and memory, and a task language to implement reasoning tasks. Nicole enables the context-sensitive asynchronous memory approach to be applied to real problems, including information retrieval in Nicole-IRIA, a information management application that uses context to recommend useful information (Francis et al. 2000), planning in Nicole-MPA, a case-based least-commitment planner that adapts multiple plans (Ram & Francis 1995) and language understanding in ISAAC (Moorman 1997), a story understanding system which uses Nicole’s retrieval system as part of its creative understanding process. Nicole and her children thus provide a testbed to evaluate the context-sensitive asynchronous memory approach.

Experiments with Nicole support the claims of the approach. Experiments with Nicole-IRIA demonstrate that a context-sensitive asynchronous memory can use feedback from browsing to improve the quality of memory retrieval, while experiments with Nicole-MPA demonstrate how information derived from reasoning can improve the quantity of retrieval. The use of Nicole’s memory in the ISAAC system demonstrates the generality of the context-sensitive asynchronous memory approach. Other experiments with Nicole-MPA demonstrate the importance of representation as a source of power for context-sensitive asynchronous memory, and further demonstrate that the core features of the experience-based agent architecture are crucial sources of power necessary to enable a reasoning task to work with and exploit a context-sensitive asynchronous memory.

In sum, these evaluations demonstrate that the context-sensitive asynchronous memory approach is a general approach to memory retrieval which can provide concrete benefits to real problems.

Read the thesis:

Context-Sensitive Asynchronous Memory

by Anthony Francis

PhD Thesis, College of Computing, Georgia Institute of Technology, Atlanta, GA, 2000
www.cc.gatech.edu/faculty/ashwin/papers/git-cc-00-01.pdf

Introspective Multistrategy Learning: On the Construction of Learning Strategies

A central problem in multistrategy learning systems is the selection and sequencing of machine learning algorithms for particular situations. This is typically done by the system designer who analyzes the learning task and implements the appropriate algorithm or sequence of algorithms for that task. We propose a solution to this problem which enables an AI system with a library of machine learning algorithms to select and sequence appropriate algorithms autonomously. Furthermore, instead of relying on the system designer or user to provide a learning goal or target concept to the learning system, our method enables the system to determine its learning goals based on analysis of its successes and failures at the performance task.

The method involves three steps: Given a performance failure, the learner examines a trace of its reasoning prior to the failure to diagnose what went wrong (blame assignment); given the resultant explanation of the reasoning failure, the learner posts explicitly represented learning goals to change its background knowledge (deciding what to learn); and given a set of learning goals, the learner uses nonlinear planning techniques to assemble a sequence of machine learning algorithms, represented as planning operators, to achieve the learning goals (learning-strategy construction). In support of these operations, we define the types of reasoning failures, a taxonomy of failure causes, a second-order formalism to represent reasoning traces, a taxonomy of learning goals that specify desired change to the background knowledge of a system, and a declarative task-formalism representation of learning algorithms.

We present the Meta-AQUA system, an implemented multistrategy learner that operates in the domain of story understanding. Extensive empirical evaluations of Meta-AQUA show that it performs significantly better in a deliberative, planful mode than in a reflexive mode in which learning goals are ablated and, furthermore, that the arbitrary ordering of learning algorithms can lead to worse performance than no learning at all. We conclude that explicit representation and sequencing of learning goals is necessary for avoiding negative interactions between learning algorithms that can lead to less effective learning.

Read the paper:

Introspective Multistrategy Learning: On the Construction of Learning Strategies

by Mike Cox, Ashwin Ram

Artificial Intelligence, 112:1-55, 1999
www.cc.gatech.edu/faculty/ashwin/papers/er-99-01.pdf