Matches in ScholarlyData for { ?s <https://w3id.org/scholarlydata/ontology/conference-ontology.owl#abstract> ?o. }
- 207 abstract "Associating meaning with data in a machine-readable format is at the core of the Semantic Web vision, and typing is one such process. Typing (assigning a class selected from schema) information can be attached to URI resources in RDF/S knowledge graphs and datasets to improve quality, reliability, and analysis. There are two types of properties: object properties, and datatype properties. Type information can be made available for object properties as their object values are URIs. Typed object properties allow richer semantic analysis compared to datatype properties, whose object values are literals. In fact, many datatype properties can be analyzed to suggest types selected from a schema similar to object properties, enabling their wider use in applications. In this paper, we propose an approach to glean types for datatype properties by processing their object values. We show the usefulness of generated types by utilizing them to group facts on the basis of their semantics in computing diversified entity summaries by extending a state-of-the-art summarization algorithm.".
- 219 abstract "Identity relations are at the foundation of the Semantic Web and the Linked Data initiative. In many instances the classical interpretation of identity is too strong for practical purposes. This is particularly the case when two entities are considered the same in some but not all contexts. Unfortunately, modeling the specific contexts in which an identity relation holds is cumbersome and, due to arbitrary reuse and the Open World Assumption, it is impossible to anticipate all contexts in which an entity will be used. We propose an alternative semantics for owl:sameAs that partitions the original relation into a hierarchy of subrelations. The subrelation to which an identity statement belongs depends on the dataset in which the statement occurs. Adding future assertions may change the subrelation to which an identity statement belongs, resulting in a context-dependent and non-monotone semantics. We show that this more complicated semantics is better able to characterize the pragmatic use of owl:sameAs as observed in Linked Open Datasets.".
- 220 abstract "Deciding which vocabulary terms to use when modeling data as Linked Open Data (LOD) is far from trivial. In this paper, we propose TermPicker: a novel approach enabling vocabulary reuse by recommending RDF types and properties based on various features of a vocabulary term. These features include the term’s popularity, whether it is from an already used vocabulary, and the so-called schema-level pattern (SLP) feature, which exploits how other data providers on the LOD cloud use RDF types and properties to describe their data. We use the machine learning approach Learning To Rank (L2R) to establish a generalized ranking model for vocabulary terms over the utilized features. The results illustrate that using the SLP feature provides an improvement of the recommendation quality by 29 − 36% considering the Mean Average Precision (MAP) and the Mean Reciprocal Rank at the first five positions (MRR@5) compared to recommendations based on solely the term’s popularity and whether it is from an already used vocabulary.".
- 222 abstract "When modeling Linked Open Data (LOD), choosing appropriate vocabulary terms to represent data entities and relations between data entities is difficult, because there are many vocabularies to choose from. Inappropriate choices lead to LOD that is difficult both to understand for humans as well as to automatically exploit by machines. We present an evaluation of approaches that try to alleviate this situation by recommending vocabulary terms based on how other data providers have used RDF classes and properties in the LOD cloud. Our user study compares the machine learning technique Learning to Rank (L2R), the classical data mining approach Association Rule mining (AR), and a baseline that does not provide any recommendations. Results show that utilizing AR, participants needed less time and less effort to model the data, which in the end resulted to be of better quality.".
- 224 abstract "The rise of Big Data Analytics has shown the utility of analyzing all aspects of a problem by bringing together disparate data sets. Efficient and accurate private record linkage algorithms are necessary to achieve this. However, records are often linked based on personally identifiable information, and protecting the privacy of individuals is critical. This paper contributes to this field by studying an important component of the private record linkage problem: linking based on names while keeping those names encrypted, both on disk and in memory. We explore the applicability, accuracy and speed of three different primary approaches to this problem (along with several variations) and compare the results to common name-matching metrics on unprotected data. While these approaches are not new, this paper provides a thorough analysis on a range of datasets containing systematically introduced flaws common to name-based data entry, such as typographical errors, optical character recognition errors, and phonetic errors.".
- 227 abstract "Emerging ontology authoring methods to add knowledge to an ontology focus on ameliorating the validation bottleneck. The verification of the newly added axiom is still one of trying and seeing what the reasoner says, because a systematic testbed for ontology authoring is missing. We sought to address this by introducing the approach of test-driven development for ontology authoring. We specify 36 generic tests, as TBox queries and TBox axioms tested through individuals, and structure their inner workings in an `open box'-way, which cover the OWL 2 DL language features. This is implemented as a Protege plugin so that one can perform a TDD test as a black box test. We evaluated the two test approaches on their performance. ".
- 230 abstract "In this paper we introduce the Semantic Topic Compass (STC) Framework, which makes use of topic priors derived from structured knowledge sources to facilitate the semantic feature grading of a topic. The proposed STC characterises ambiguity in a topic's feature space. In this paper we investigate the impact of the use of ambiguous lexicons in a topic classification task. Our findings show promising results in understanding the impact of ambiguity in feature relevance assessment for topic classification. ".
- 238 abstract "Finding relevant resources on the Semantic Web today is a dirty job: there is no centralized lookup service and the support for natural language lookup is limited. In this paper, we present LOTUS: Linked Open Text UnleaShed, a natural language entry point to a massive subset of today's Linked Open Data Cloud. While a wide array of matching and ranking algorithms have been studied, there is no ultimate combination of matching and ranking that will work for every use case. LOTUS recognizes the case-dependent nature of resource retrieval by allowing users to choose from a wide palette of well-known matching and ranking algorithms. LOTUS is an adaptive toolkit that allows user to easily construct the form of resource retrieval that suits her use case best. In this paper, we explain the LOTUS approach, its implementation and the functionality it provides. We also demonstrate the ease with which LOTUS allows Linked Data to be queried at an unprecedented scale in different concrete and domain-specific scenarios. Finally we demonstrate the scalability of LOTUS with respect to the LOD Laundromat data, the biggest collection of easily accessible Linked Data currently available.".
- 241 abstract "Over the years, Twitter has become one of the largest communication platforms providing key data to various applications such as brand monitoring, trend detection, among others. Natural language understanding from tweets is particularly difficult due to prevalent short length and unconventional grammar. Entity linking associates text to knowledge bases in order to provide unambiguous references and additional context to help with understanding. State-of-the-art techniques have focused on linking explicitly mentioned entities on tweets with reasonable success. However, we argue that in addition to explicit mentions -- i.e. `You guys have no idea how wet the movie Gravity makes me' -- entities (movie Gravity) can also be mentioned implicitly -- i.e. `This new space movie is crazy. you must watch it!'. This paper introduces the problem of implicit entity linking in tweets. We propose an approach that models the entities by exploiting their factual and contextual knowledge. We demonstrate how to use these models to perform implicit entity linking on a ground truth dataset with 400 tweets from two domains, namely, movie and book. Specifically, we show: 1) the value of performing implicit entity linking along with explicit entity linking task, and the importance of exploiting contextual knowledge of an entity in linking their implicit mentions on tweets.".
- 243 abstract "An increasing number of everyday tasks involve a mixture of human actions and machine computation. This paper presents the first framework that allows non-programmer users to create and execute workflows where each task can be completed by a human or a machine. In this framework, humans and machines interact through a shared knowledge base which is both human and machine understandable. This knowledge base is based on the PROHOW Linked Data vocabulary that can represent human instructions and link them to machine functionalities. Our hypothesis is that non-programmer users can describe how to achieve certain tasks at a level of abstraction which is both human and machine understandable. This paper presents the PROHOW vocabulary and describes its usage within the proposed framework. We substantiate our claim with a concrete implementation of our framework and by experimental evidence.".
- 251 abstract "Querying and reasoning over RDF streams are two increasingly relevant areas in the broader scope of processing structured data on the Web. While RDF Stream Processing (RSP) has focused so far on extending SPARQL for continuous query and event processing, stream reasoning has concentrated on ontology evolution and incremental ma- terialization. In this paper we propose a di erent approach for querying RDF streams over ontologies, based on the combination of query rewriting and stream processing. We show that it is possible to rewrite continuous queries over streams of RDF data, while maintaining e - ciency for a wide range of scenarios. We provide a detailed description of our approach, as well as an implementation, StreamQR, which is based on the kyrie rewriter, and can be coupled with a native RSP engine, namely CQELS. Finally, we show empirical evidence of the performance of StreamQR in a series of experiments based on the SRBench query set.".
- 253 abstract "Benchmarks like LSBench, SRBench, CSRBench and, more recently, CityBench satisfy the growing need of shared datasets, ontologies and queries practice to evaluate window-based RDF Stream Processing (RSP) engines. However, no clear winner emerges out of the evaluation. In this paper, we claim that the RSP community needs to adopt a Systematic Comparative Research Approach (SCRA) if it wants to move a step forward. To this end, we propose a framework that enables SCRA for window based RSP engines. The contributions of this paper are: (i) the requirements to satisfy for tools that aim at enabling SCRA; (ii) the architecture of a facility to design and execute experiment guaranteeing repeatability, reproducibility and comparability; (iii) Heaven - a proof of concept implementation of such architecture that we released as open source -; (iv) two RSP engine implementations, also open source, that we propose as baselines for the comparative research (i.e., they can serve as terms of comparison in future works). We evaluate Heaven effectiveness using the baselines by: (i) showing that top-down hypothesis verification is not straight forward even in controlled conditions and (ii) providing examples of bottom-up comparative analysis.".
- 38 abstract "With recent advances in information extraction techniques, various large-scale knowledge bases covering a broad range of knowledge have become publicly available. As no single knowledge base covers all information, many applications require access to integrated knowledge from multiple knowledge bases. Achieving this, however, is challenging due to their differences in knowledge representation.To address this problem, this paper proposes to use linguistic frames as a common representation and maps heterogeneous knowledge bases to the FrameBase schema, which is formed by a large inventory of linguistic frames. We develop several methods to create complex mappings from external knowledge bases into this schema, using text similarity measures, machine learning and different heuristics. We test them with different widely used large-scale knowledge bases, YAGO2s, Freebase and WikiData. The resulting integrated knowledge can then be queried in a homogeneous way.".
- 40 abstract "With the increasing quantity and diversity of publicly available Web datasets, most notably, Linked Open Data, recommending datasets, which meet specific criteria, has become an increasingly important, yet challenging problem. Dataset recommendation is an important task when addressing issues such as entity retrieval, semantic search and, particularly, data linking, where one aims to identify datasets, which are likely to contain linking candidates. While an understanding of the nature of the content of specific datasets is a crucial prerequisite, we adopt the notion of dataset profiles, where a dataset is characterized through a set of schema concept labels that best describe it. Alternatively, by retrieving the textual descriptions of each of these labels, we can map the profiles to text documents. We introduce a dataset recommendation approach to identify linking candidates. The method is based on the presence of schema overlap between datasets by computing a semantico-frequential similarity between profiles and a ranking criterium based on the tf-idf cosine similarity. Our experiments, applied to all available Linked Datasets in the Linked Open Data (LOD) cloud, show that our method achieve an average precision up to 53% for a recall of 100%. An additional contribution of this work is the mapping returned for the schema concepts between datasets which are particularly useful in the linking step.".
- 45 abstract "Powered by Semantic Web technologies the Linked Data paradigm aims at weaving a globally interconnected graph of raw data that transforms the ways we publish, retrieve, share, reuse, and integrate data from a variety of distributed and heterogeneous sources. In practice, however, this vision faces substantial challenges with respect to data quality, coverage, and longevity, the amount of background knowledge required to query distant data, the reproducibility of query results and their derived (scientific) findings, and the lack of computational capabilities required for many tasks. One key issues underlying these challenges is the trade-off between storing data and computing them. Intuitively, data that is derived from already stored data, changes frequently in space and time, or is the result of some workflow or procedure, should be compute. However, this functionality is not readily available on the Linked Data cloud and its technology stack. In this work, we introduce a proxy that can transparently run on top of arbitrary SPARQL endpoints and enables the on-demand computation of Linked Data together with the provenance information required to understand how they were derived. While our work can be generalized to multiple domains, we focus on two geographic use case to showcase the proxy's capabilities.".
- 5 abstract "Ontology reasoning is typically a computationally intensive operation. While soundness and completeness of results is required in some use cases, for many others, a sensible trade-off between computation efforts and correctness of results makes more sense. In this paper, we show that it is possible to approximate the results delivered by a reasoner by observing an actual reasoner and training a machine learning model which approximates the behavior of that reasoner. On four different datasets, we show that such learned models constantly achieve an accuracy above 95% at less than 2% of the runtime of a reasoner, using a decision tree with no more than 20 inner nodes. For example, this allows for checking the consistency of all relational assertions in DBpedia in less than 90 minutes, compared to 15 days required by a state of the art ontology reasoner. ".
- 51 abstract "In this work, we envision a publish/subscribe ontology system that is able to index large numbers of expressive continuous queries and filter them against RDF data that arrive in a streaming fashion. To this end, we propose a SPARQL extension that supports the creation of full-text continuous queries and propose a family of main-memory query indexing algorithms which perform matching at low complexity and minimal filtering time. We experimentally compare our approach against a state-of-the-art competitor (extended to handle indexing of full-text queries) both on structural and full-text tasks using real-world data. Our approach proves two orders of magnitude faster than the competitor in all types of filtering tasks.".
- 7 abstract "In the Resource Description Framework, literals are composed of a UNICODE string (the lexical form), a datatype IRI, and optionally, when the datatype IRI is rdf:langString, a language tag. Any IRI can take the place of a datatype IRI, but the specification only defines the precise meaning of a literal when the datatype IRI is among a predefined subset. Custom datatypes have reported use on the Web of Data, and show some advantages in representing some classical structures. Yet, their support by RDF processors is rare and implementation specific. In this paper, we first present the minimal set of functions that should be defined in order to make a custom datatype usable in query answering and reasoning. Based on this, we discuss solutions that would enable: (i) data publishers to publish the definition of arbitrary custom datatypes on the Web, and (ii) generic RDF processor or SPARQL query engine to discover custom datatypes on-the-fly, and to perform operations on them accordingly. Finally, we detail a concrete solution that targets arbitrarily complex custom datatypes, we overview its implementation in Jena and ARQ, and we report the results of an experiment on a real world DBpedia use case. ".
- 81 abstract "Links between knowledge bases build the backbone of the Linked Data Web. Several time-efficient algorithms have hence been developed for computing links between knowledge bases. However, these approaches pay little attention to the fact that very large datasets cannot be held in the main memory of most computing devices. In this paper, we address this research gap by presenting a generic memory management approach that can by combined with any blocking- or filtering-based algorithm for Link Discovery. We show that the problem at hand is a variation of the Traveling Salesman Problem and is thus NP-complete. We thus provide efficient best-effort graph-based algorithms that allow scheduling link discovery tasks efficiently. We combine these algorithms with caching approaches that allow for loading required data from massive storage. Our evaluation on real data shows that our approach allows computing links between large amounts of resources efficiently even when using commodity hardware.".
- 89 abstract "Recent advances in Web-information extraction have led to the creation of several large Knowledge Bases (KBs). Querying these KBs often results in empty answers that do not serve the users' needs. Relaxation of the failing queries is one of the cooperative techniques used to retrieve alternative results. Most of the previous work on RDF query relaxation compute a set of relaxed queries and execute them in a similarity-based ranking order. Thus, these approaches relax an RDF query without knowing its Failure Causes (FCs). In this paper, we study the interest of identifying these FCs to speed up the query relaxation process. We propose three relaxation strategies based on various information levels about the FCs of the user query and of its relaxed queries as well. A set of experiments conducted on the LUBM benchmark show the interest of our proposal in comparison with a state-of-the-art algorithm.".
- 1 abstract "The Linked Data initiative gained momentum inside as well as outside of theresearch community. Thus, it is already an accepted research issue to investigate usage mining in the context ofthe Web of Data from various perspectives. We are currently working onan approach that applies such usage mining methods and analysis to support ontology and datasetmaintenance tasks. This paper presents one part of this work, namely a methodto detect errors or weaknesses within ontologies used for Linked Data populationbased on statistics and network visualizations. We contribute a detailed description of a log file preprocessing algorithm for Web of Data endpoints, a set of statistical measures that help to visualize different usage aspects, and an examplary analysis of one of the most prominent Linked Data set -- DBpedia -- aimed to show the feasibility and the potential of our approach.".
- 10 abstract "As ontologies evolve, it becomes important to be able to dis-cover how they have changed over time. In the recent past there havebeen a few very useful tools based on the Manchester OWL API thataddress this issue. However, these new tools do not take into accounthow to align entities in two ontologies when the names of the entitieshave changed. In this case, we need to be able to discover the alignmentbetween the entities in the two ontologies before can match up the struc-tures in the two ontologies. In this paper, we describe a highly optimizedpluggable difference engine that searches for alignments between entitiesin different ontology versions and applies those alignments to display thedifferences in the ontologies. We discuss our experiences applying thetools to a selection of ontologies from the BioPortal ontology reposi-tory, including the performance and accuracy of the tool.".
- 11 abstract "The analysis of changes between OWL ontologies (in the form of a diff) is an important service for ontology engineering. A purely syntactic analysis of changes is insufficient to distinguish between changes that have logical impact and those that do not. The current state of the art in semantic diffing ignores logically ineffectual changes and lacks any further characterisation of even significant changes. We present and demonstrate a diff method based on an exhaustive categorisation of effectual and ineffectual changes between ontologies. In order to verify the applicability of our approach we apply it to 88 versions of the National Cancer Institute (NCI) Thesaurus (NCIt), and 5 versions of SNOMED CT, demonstrating that all categories are realized throughout the corpus. Based on the outcome of these studies we argue that the devised categorisation of changes is helpful for ontology engineers and their understanding of changes carried out between ontologies.".
- 12 abstract "Ontologies can be a powerful tool for structuring knowledge, and they are currently the subject of extensive research. Updating the contents of an ontology or improving its interoperability with other ontologies is an important but difficult process. In this paper, we focus on the presence of vague concepts, which are pervasive in natural language, within the framework of formal ontologies. We will adopt a framework in which vagueness is captured via numerical restrictions that can be automatically adjusted. Since updating vague concepts, either through ontology alignment or ontology evolution, can lead to inconsistent sets of axioms, we define and implement a method to detecting and repairing such inconsistencies in a local fashion.".
- 3 abstract "Knowing about language evolution can significantly help to reveal lost information and help access documents containing language that has long since been forgotten. In this position paper we will report on our methods for finding word senses and show how these can be used to reveal important information about their evolution over time. We discuss the weaknesses of current approaches and outline future work to overcome these weaknesses. ".
- 5 abstract "There have been several attempts to make RDBMS and RDFstores inter-operate. The most popular one, D2RQ, has explored onedirection i.e. to look at RDBMS through RDF lenses. In this paper wepresent RETRO, which investigates the reverse direction i.e. to look atRDF through Relational lenses. RETRO generates a relational schemafrom an RDF store, enabling a user to query RDF data using SQL.A significant advantage of this direction in-addition to interoperabilityis that it makes numerous relational tools developed over past severaldecades, available to the RDF stores. In order to provide interoperabilitybetween these two DB systems one needs to resolve the heterogeneitybetween their respective data models and include schema mapping, datamapping and query mapping in the transformation process [1]. However,like D2RQ, RETRO chooses not to physically transform the data anddeals only with schema mapping and query mapping. RETRO’s schemamapping derives a domain specific relational schema from RDF dataand its query mapping transforms an SQL query over the schema into aprovably equivalent SPARQL query, which in-turn is executed upon theRDF store. Since RETRO is a read-only framework, its query mappinguses only a relevant and relationally complete subset of SQL. A proofof correctness of this transformation is given based on compositionalsemantics of the two query languages.".
- 7 abstract "Ontologies can support a variety of purposes, ranging fromcapturing conceptual knowledge to the organisation of digital contentand information. However, information systems are always subject tochange and ontology change management can pose challenges. We investigateontology change representation and discovery of change patterns.Ontology changes are formalised as graph-based change logs. We useattributed graphs, which are typed over a generic graph with node andedge attribution.We analyse ontology change logs, represented as graphs,and identify frequent change sequences. Such sequences are applied as a reference in order to discover reusable, often domain-specific and usage driven change patterns. We describe the pattern discovery algorithms and measure their performance using experimental results.".
- 9 abstract "Sequential ontology debugging aims to the efficient discrimination between diagnoses. By querying additional information the debugger can gradually reduce the number of diagnoses to be considered by the user. The selection of the best queries is of central importance for minimizing diagnosis costs. If prior fault probabilities are available, the best results are achieved by entropy based selection methods.However, given some weakly justified priors these methods bravely suggest suboptimal queries. In such a case, it is more efficient to use a no-risk method which prefers queries that eliminate 50% of diagnoses independently of any fault probabilities. However, choosing the appropriate method in advance is impossible because the quality of given priors cannot be assessed before additional information is queried.In this paper we propose a method which combines advantages of both approaches. On the one hand the method takes into account available fault probabilities and the user".
- 11 abstract "In the last decade much research has been conducted on analyzing human eye and gaze movements using eye tracking technology, not only in the fields of neuroscience, psychology and marketing, but also in the field of human computer interaction. However, no flexbile framework exist to integrate eye tracking directly into web applications to easily create and test new interaction concepts. We have created a JavaScript library based on the latest HTML5 Web technology and the jQuery library to close this gap. Facilitated by HTML5 WebSocket, the browser directly receives gaze input samples from an eye tracker, generating events that are similar to those of a mouse input device. Events like gazeOver/-Out or fixationStart/-End can be attached to any HTML element in the DOM tree. New custom events derived from the eye tracking data, e.g.blink or read, can easily be added. Using this library we have successfully implemented a number of Web applications, allowing the user to interact with their eyes. This paper also describes our gaze enabled Web-based eLearning environment. Our JavaScript library is used within the eLearning environment to capture and interpret eye gaze events for the purpose to support users in the acquisition of new knowledge.".
- 12 abstract "Emerging Technologies for Education affect a broad spectrum of schools, universities and other educational institutions, academic disciplines, pedagogy, administration, innovation, and, most importantly, the people involved, viz. students, teachers, administrators and, at a remove, parents and society at large. It is therefore necessary to define at the outset how much of this spectrum one wants to cover. This paper reports on efforts and experience of a group of academics in computing in a university who have used the Web as an emerging technology, from 1996 onwards, specifically in terms of how it affected the delivery of education at tertiary level. The paper describes initiatives that led to changes in the administration and examination techniques at an individual level at first and then, in successive as well as parallel steps, in administrative procedures, pedagogy and careful attention to emerging technologies and student cohorts across the university. There are several lessons from this engagement. In chronological order, the first conclusion is that introduction of online delivery of educational material by a few academics generated pressure on the others to emulate the effort, leading to demands from them for easy procedures to meet the demands from both students and administration. Next came changes to curriculum to introduce the emerging technologies to the undergraduate students in computing, followed by the recognition that undergraduate students were ill-prepared to analyse and understand the implications of the emerging technologies. This led to the introduction of a specially designed postgraduate programme in IT. Pedagogically, the emphasis shifted from quizzes and examinations to project work over a semester, at times continuing onto another semester. New technologies were introduced as assignments, short projects and group projects. Existing and new Web sites and Web applications were tracked along with the administrative policies within the university to add to the teaching and learning experience. Over time, behaviour of student cohorts has also changed from being predominantly innovators and experimenters to consumers in the main, waiting to be shown how the new technologies could be used to create new and/or better applications.".
- 13 abstract "Extended cognition is now a reality with rise of social web. Smart devices and emerging collective intelligence is aiding us to take part in tasks much bigger than we can naturally handle. In the age of digital natives, learning is the most affected process by this phenomenon, and has created a void in this space to rethink the model to suit the generation of web. Current educational model is not future proof as its creating more autonomous problem solvers, while future demands high caliber people to collaborate on interdisciplinary problems with potential global impact. Primary motto of this learning model is to develop critical thinking andcontinuous learning among individuals. Can such process be engineered in the first place? If ones goal is to attain the formal derivatives, what are the possible ways to realize it? Current paper discusses along with the generic web learning trends, a model based on Rhizomatic learning and contextual relations generated from similarity sets of social networks. This unique approach emphasizes more on distance among the similar sets to promote maximum diversity in the learning flows. Also leverages our earlier work, a feedback framework designed to judge diverse facets of a personality from interactions on the web.".
- 2 abstract "Energy sustainability is prevalent in political and popular rhetoric and yet energy consumption is rising. Teenagers are an important category of future energy consumers, but little is known of their conceptions about energy, energy saving, and energy related problems. We report on a study with a group of teenagers that explored their conceptions about¬ energy and their skills in finding information about their personal energy consumption. In this paper we focus our discussion on the challenges in using Web technologies to support learning about complex real world issues like energy consumption. An initial analysis of the data indicates that teenagers struggle to grasp the complexity of problems around energy and to search for information on examples of their personal energy consumption. We highlight that, to fulfil the learning potential of Web technologies, educators must provide support both in terms of learners’ initial conceptual understanding of the learning topic and the process of searching for information. Motivation is also critical if learners are to engage in the time intense process of searching for information and creating content. These findings have implications for the design of technology enhanced learning experiences that build on young people’s Web technology skills to support learning about complex real world issues.".
- 3 abstract "We investigate how the three technologies, social media, mobile / pervasive learning and semantic web, may enhance Inquiry-Based Science Teaching (IBST) approaches and digital literacy. IBST may be defined by engaging students in: i) authentic and problem-based activities, ii) experimental procedures, iii) self regulated learning sequences, iv) discursive argumentation and communication with peers. We analyze the benefits of each technology and their combination. From an existing IBST learning scenario, we propose an innovative one based on the convergence of the three following technologies: social media, mobile / pervasive learning and semantic web.".
- 428 abstract "Many organizations face every day the problem of effectively managing their contacts (customers, suppliers, partners, etc.), in terms of communication, clustering, networking, analysis, and so on. Our company decided to cope with this issue by gathering the requirements for Contact Management and by designing and developing a prototype, called GeCo, to fit Cefriel needs. During the development of this application, which ran in parallel with some research projects dealing with Semantic Web technologies, we recognized that the addition of some "semantics", both in the data modeling and in the tool design, would help a lot in solving the open issues for the general problem of Contact Management. In this paper we summarize the main criticalities in managing contacts and we suggest how Semantic Web technologies can contribute to their successful solution.".
- 463 abstract "This paper proposes a methodology for constructing Semantic Campus, a Semantic Web application that represents the social network of the academics in the university, King Mongkut’s Institute of Technology North Bangkok. Semantic Campus is constructed by extracting data that is available on the web site of the university. The extracted data is analyzed with respect to its association to terms defined in an ontology and associations between people in the university to reveal whether or not one academic knows another. We also discuss relation analysis that considers direct and indirect association of the campus-based resources (i.e. knowledge about the people) contained in Semantic Campus. Such analysis can be used, for example, to find specific experts in the university and research interests shared by a number of academics.".
- 500 abstract "When we want to find experts on the Web we might want to search where the knowledge is created by the users. One of such knowledge repository is Wikipedia. People expertises are described in Wikipedia pages and also the Wikipedia users can be considered experts on the topics they produce content on. In this paper we propose algorithms to find experts in Wikipedia. The two different approaches are finding experts in the Wikipedia content or among the Wikipedia users. We also use semantics from WordNet and Yago in order to disambiguate expertise topics and to improve the retrieval effectiveness. In the end, we show how our methodology can be implemented in a system in order to improve the expert retrieval effectiveness.".
- 507 abstract "Research collaborations are always encouraged, as they often yield good results. However, the researcher network contains massive amounts of experts in various disciplines and it is difficult for the individual researcher to decide which experts will match his own expertise best. As a result, collaboration outcomes are often uncertain and research teams are poorly organized. We propose a method for building link predictors in networks, where nodes can represent researchers and links - collaborations. In this case, predictors might offer good suggestions for future collaborations. We test our method on a researcher co-authorship network and obtain link predictors of encouraging accuracy. This leads us to believe our method could be useful in building and maintaining strong research teams. It could also help with choosing vocabulary for expert description, since link predictors contain implicit information about which structural attributes of the network are important with respect to the link prediction problem.".
- 515 abstract "This paper shows a method for finding topic-centric experts from open access metadata and full text documents. Topic-centric information including experts is served on OntoFrame, which is a Semantic Web-based academic research information service supporting R&D activities. URI scheme-based OntoFrame provides three entity pages: topic, person, and event. ‘Persons by Topic’ in topic page lists up topic-centric identified experts. SPARQL query is used to re-trieve them from RDF triple store through backward chaining. We gathered CiteSeer open access metadata and full text documents with the amount of about 110,000 papers. Using about 160,000 abundant topics, OntoFrame now serves topic-centric identified experts and relevant information acquired by full text analysis.".
- 501 abstract "A lot of functionality is needed when an application, such as a museum cataloguing system, is extended with semantic capabilities, for example ontological indexing functionality or multi-facet search. To avoid duplicate work and to enable easy and cost-efficient integration of information systems with the Semantic Web, we propose a web widget approach. Here, data sources are combined with functionality into ready-to-use software components that allow adding semantic functionality to systems with just a few lines of code. As a proof of the concept, we present a collection of general semantic web widgets and case applications that use them, such as the ontology server ONKI, the annotation editor SAHA and the culture portal CultureSampo.".
- 525 abstract "Data Integration refers to the problem of combining data residing at autonomous and heterogeneous sources, and providing users with a unified global view. Ontology-based integration solutions have been advocated but for the case to be made for real deployment of such solutions, the integration effort and performance needs to be characterized. In this paper, we measure the performance of a generalised ontology based integration system using the THALIA integration benchmark. The ontology based integration solution is used to integrate data dynamically across a real telecommunications value chain. An extension of the THALIA benchmark, to take account of the integration effort required, is introduced. We identify the issues impacting the ontology based integration approach and propose further experiments.".
- 529 abstract "Nowadays enterprises request information technologies that leverage structured and unstructured information for providing a single integrated view of business problems in order to foster better business process management and decision making. The growing interest in semantic technologies is due to the limitation of existing enterprise information technologies to answer these new challenging needs. Semantic Web Technologies (SWT), the current open standard approaches to semantic technologies based on semantic web languages, provide some interesting answers to novel enterprise needs by allowing to use domain knowledge within applications. However, SWT aren't well suited for enterprise domain because of some drawbacks and a lack of compatibility with enterprise-class applications. This paper presents the new Semantic Enterprise Technologies (SET) paradigm founded on the idea of Semantic Models that are executable, flexible and agile representation of domain knowledge. Semantic Models are expressed by means of the Codex Language obtained combining Disjunctive Logic Programming (Datalog plus disjunction) and Attribute Grammars both extended by object-oriented and two-dimensional capabilities. Semantic Models enable to exploit domain knowledge for managing both structured and unstructured information. Since the Codex Language derives from the database field, it allows SET to provide advanced semantic capabilities well suited for enterprises. Differences and interoperability issue between SET and SWT are briefly discussed in the paper that shows, also the SET Reference Architecture (SETA), an application example and the business value of SET.".
- 531 abstract "The use of Semantic Grid architecture eases the development of complex, flexible applications, in which several organisations are involved and where resources of diverse nature (data and computing elements) are shared. This is the situation in the Space domain, with an extensive and heterogeneous network of facilities and institutions. There is a strong need to share both data and computational resources for complex processing tasks. One such is monitoring and data analysis for Satellite Missions and this paper presents the Satellite Mission Grid, built in the OntoGrid project as an alternative to the current systems used. Flexibility, scalability, interoperability, extensibility and efficient development were the main advantages found in using a common framework for data sharing and creating a Semantic Data Grid.".
- 533 abstract "As the semantic web vision continues to proliferate a gap still remains in the full scale adoption of such technologies. The exact reasons for this continue to be the subject of ongoing debate, however, it is likely the emergence of reproducible infrastructure and deployments will expedite its adoption. We illustrate the recognizable added value to life science researchers gained through the convergence of existing and customized semantic web technologies (content acquisition pipelines supplying legacy unstructured texts, natural language processing, OWL-DL ontology development and instantiation, reasoning over A-boxes using a visual query tool). The resulting platform allows lipidomic researchers to rapidly navigate large volumes of full-text scientific documents according to recognizable lipid nomenclature, hierarchies and classifications. Specifically we have enabled searches for sentences describing lipid-protein and lipid-disease interactions.".
- 534 abstract "Employment Services (ESs) are becoming more and more important for Public Administrations where their social implications on sustainability, workforce mobility and equal opportunities play a fundamental strategic importance for any central or local Government. The EU SEEMP project aims at improving facilitate workers mobility in Europe. Ontologies are used to model descriptions of job offers and curricula; and for facilitating the process of exchanging job offer data and CV data between ES. In this paper we present the methodological approach we followed for reusing existing human resources management standards in the SEEMP project, in order to build a common “language” called Reference Ontology.".
- 1 abstract "In the last few years, the concept of network virtualization has gained a lot of attention both from industry and research projects. This paper evaluates the potential of network virtualization from an operator’s perspective, with the short-term goal of optimizing service delivery and rollout, and on a longer term as an enabler of technology integration and migration. Based on possible scenarios for implementing and using network virtualization, new business roles and models are examined. Open issues and topics for further evaluation are identified. In summary, the objective is to identify the challenges but also new opportunities for telecom operators raised by network virtualization.".
- 2 abstract "This paper examines the impact of the cloud computing paradigm on enterprise IT structures. The first section of the paper defines the term cloud computing with the relevant aspects for this paper. In the second chapter the paper shows the potential of cloud computing for the enterprise IT. In the last section the paper will briefly reflect how the future development of infrastructure technologies will create more potential for enterprise IT cloud cases in the future.".
- 1 abstract "Malicious software (malware) infects large numbers of computers around the world. This malware can be used to promote unwanted products, disseminate offensive content, or provide unauthorized access to personal and financial information. Until recently mobile networks have been relatively isolated from the Internet, so there has been little need to protect them against Botnets. Mobile networks are now well integrated with the internet, so threats on the internet such as Botnets have started to migrate onto mobile networks. Botnets on mobile devices will probably appear very soon, there are already signs that this is happening. This paper studies the potential threat of Botnets based on mobile networks, and proposes the use of computational intelligence techniques to detect Botnets. We then simulate anomaly detection followed by an interpretation of the simulated values.".
- 2 abstract "The combination of semantic technology and Web services in form of `Semantic Web Services' has until now been oriented towards extension of the {WS-*} stack with ontology-based descriptions. The same time, there is a strong movement away from this stack --- for which the `Web' part is little more than branding --- towards RESTful services. The Linked Open Data initiative is a keen adopter of this approach and exposes many datasets via SPARQL endpoints and RESTful services. Our developing approach of `Linked Open Services', whose current state is described in this paper, accommodates such Linked Data endpoints and general RESTful services alongside {WS-*} stack-based services with descriptions based on RDF and SPARQL. This capitalises on the Linked Data Cloud and makes service description and comprehension more easy and direct to the growing Linked Data community. Along the way, we show how the existing link between service messaging and the semantic viewpoint, commonly called `lifting and lowering', is usually unduly restricted to ontology-based classification and misses how the effect of a service contributes to the knowledge of its consumer. Our SPARQL-based approach helps also in the composition of services as knowledge-centric processes, and encourages the development and exposure of services that communicate RDF.".
- 3 abstract "We investigate services giving users an adequate insight on his or her energy consumption habits in order to optimize it in the long run. The explored energy awareness services are addressed to inhabitants of smart homes, equipped with smart meters, advanced communication facilities, sensors and actuators. To analyze the potential of such services, a game at a social network Facebook has been designed and implemented, and the information about players' responses and interactions within the game environment has been collected and analyzed. The players have had their virtual home energy usage visualized in different ways, and had to optimize the energy consumption basing on their own perceptions of the consumption information. Evaluations reveal, in particular, that users are specifically responsive to information shown as a real-time graph and as costs in Euro, and are able to produce and share with each other policies for managing their smart home environments.".
- 4 abstract "The Future Internet, has as its vision, the development of improved features and usability for services, applications and content. In many cases, services can be provided automatically through the use of monitors or sensors. This means web generated sensor data becoming available not only to the companies that own the sensors but also to the domain users who generate the data and to information and knowledge workers who harvest the output. The goal is improving the service through better usage of the information provided by the service. Applications and services vary from climate, traffic, health and sports event monitoring. In this paper, we present the WSW system that harvests web sensor data to provide additional and, in some cases, more accurate information using an analysis of both live and warehoused information.".
- 5 abstract "Business applications are facing an increasing demand for being delivered as on-demand services. Service Level Agreements (SLAs) are a common way for specifying the exact conditions how services are delivered, both for business- and IT-based services. At the same time, the paradigm of service-orientation is expanded to different domains such as business and IT-based services. Services are composed out of other services across domains. Consequently, SLAs have to managed across such service hierarchies. In this paper, we present the lessons learned from applying a generic, multi-layer SLA management framework to the context of on-demand business applications. We explain the implemented architecture and show how a hierarchy of services can be managed via SLAs. We explain in detail how SLAs can be established for different layers and how SLA terms are translated across the service hierarchy.".
- 1 abstract "The G-Lab project aims to investigate concepts and technologies for future networks in a practical manner. Thus G-Lab consists of two major fields of activities: research studies of future network components and the design and setup of experimental facilities. Both is controlled by the same community to ensure, that the experimental facility fits to the demand of researchers. Researchers gain access to virtualized resources or may gain exclusive access to resource if necessary. We present the current setup of the experimental facility, describing the available hardware, management of the platform, the utilization of the Planet-Lab software and the user management.".
- 2 abstract "In order for a Future Internet architecture to be globally deployed, it must ensure that existing applications will continue to operate efficiently on top of it. As part of the Publish Subscribe Internet Routing Paradigm (PSIRP) project, we have explored various options for making endpoint centric applications based on the Sockets Application Programming Interface (API) compatible with the information centric PSIRP prototype implementation. We developed an emulator that mediates between the client/server socket calls and the publish/subscribe PSIRP calls, transforming the exchange of packets to distribution of publications. To assess the overhead of our emulator, we measure the execution time of a simple file transfer application in native socket mode, in emulated socket mode and in native publish/subscribe mode.".
- 3 abstract "This paper discusses the construction of a Global Semantic Graph (GSG) to support future information- and collaboration-centric applications and services. The GSG is a publish/subscribe (pub/sub) based architecture that supports publication of tuples and subscriptions with standing graph queries. We believe that an implementation of an efficient pattern matching algorithm such as Rete on top of a distributed environment might serve as a possible substrate for GSG's pub/sub facility. Rete operates on loosely coupled alpha, beta and join nodes and therefore has been chosen by us for implementation in a distributed setting. In this paper, we propose a way to perform Rete's pattern matching over a DHT-based Structured P2P network to provide a scalable content-based publish/subscribe service.".
- 4 abstract "The Future Internet faces many challenges for new Internet architectures, protocols and services requiring early experimentation and testing in large-scale environments. Furthermore, the complexity of the software and technologies running on the Future Internet will require more innovative and different approaches to the testing process. This paper examines these requirements and challenges and demonstrates how they are dealt with in the Telecommunications Software and Systems Group using a case study of the Future Internet project; PERIMETER. Details of the testing and experimental methodologies and the role of testbed activities involved in this Quality of Experience network mobility project to develop and adapt to the needs of the Future Internet are provided. The paper concludes with observations of the further developments and innovations required to achieve an end-to-end solution for experimental testing in PERIMETER, and more generally for other Future Internet projects.".
- 5 abstract "The Future Internet faces many challenges for new Internet architectures, protocols and services requiring early experimentation and testing in large-scale environments. Furthermore, the complexity of the software and technologies running on the Future Internet will require more innovative and different approaches to the testing process. This paper examines these requirements and challenges and demonstrates how they are dealt with in the Telecommunications Software and Systems Group using a case study of the Future Internet project; PERIMETER. Details of the testing and experimental methodologies and the role of testbed activities involved in this Quality of Experience network mobility project to develop and adapt to the needs of the Future Internet are provided. The paper concludes with observations of the further developments and innovations required to achieve an end-to-end solution for experimental testing in PERIMETER, and more generally for other Future Internet projects.".
- 6 abstract "Publish/Subscribe is often regarded as a promising paradigm for Future Internet architectures. Its information oriented nature and its particular security features have stimulated current research efforts which aim at applying publish/subscribe principles to a clean-slate Internet architecture. One of the core components of publish/subscribe architectures is the rendezvous network. Any security failure that a rendezvous network may face will probably jeopardize the operation of the whole (inter-)network. In this paper we highlight security requirements and potential security issues of rendezvous networks and we present security solutions that can be applied in order to shield them.".
- 1 abstract "Querying a Wiki must be simple enough for beginning users, yet powerful enough to accommodate experienced users. To this end, the keyword-based KiWi query language (KWQL) supports queries ranging from simple lists of keywords to expressive rules for selecting and reshaping Wiki (meta-)data. In this demo, we showcase visKWQL, a visual interface for the KWQL language aimed at supporting users in the query construction process. visKWQL and its editor are described, and their functionality is illustrated using example queries. The editor provides guidance throughout the query construction process through hints, warnings and highlighting of syntactic errors. The editor enables round-tripping between the twin languages KWQL and visKWQL, mean- ing that users can switch freely between the textual and visual form when constructing or editing a query. It is implemented using HTML, JavaScript, and CSS, and can thus be used in (almost) any web browser without any additional software.".
- 10 abstract "The NoTube European project looks at creating the future of TV, in which the TV user will be placed back in the driver’s seat by having a personalized TV experience with rich interaction possibilities. For this, new technologies like Linked Data, Semantic Web and Social Web data analysis are applied to the TV domain.".
- 11 abstract "We present the Information Workbench, an application for interacting with the Web of data. The Information Workbench man- ages large amounts of structured and unstructured information, which may be imported and integrated from existing sources, but also allows end users to annotate, complete and update information in a collaborative way. New paradigms for accessing information include hybrid search across the structured and unstructured data, keyword search combined with facetted search, as well as semantic query completion and interpretation, which assists the user in expressing complex information needs by an automated translation of keyword queries into hybrid queries. A Living UI based on widgets for the interaction with the data enables a homogeneous, seamless, continuous and personal experience.".
- 2 abstract "Web crawlers are increasingly used for focused tasks such as the extraction of data from Wikipedia or the analysis of social networks like last.fm. In these cases, pages are far more uniformly structured than in the general Web and thus crawlers can use the structure of Web pages for more precise data extraction and more expressive analysis. In this demonstration, we present a focused, structure-based crawler generator, the "Not so Creepy Crawler" (nc2). What sets nc2 apart, is that all analysis and decision tasks of the crawling process are delegated to an (arbitrary) XML query engine such as XQuery or Xcerpt. Customizing crawlers just means writing (declarative) XML queries that can access the currently crawled document as well as the metadata of the crawl process. We identify four types of queries that together suffice to realize a wide variety of focused crawlers.".
- 3 abstract "Linked Open Services are a principled attempt to guide the creation and exposure of online services, in order to enable productive use of, and contribution to, Linked Data. As a technical basis, they communicate RDF directly via HTTP-based interaction, and their contribution to the consumer’s knowledge, in terms of Linked Data, is described using SPARQL-based constructs. In this demo we will show some of our latest results in the production of Linked Open Services and the applications in which they can be taken up.".
- 4 abstract "ERP B3 is a SLA@SOI framework based solution for Service Level Agreement driven, on demand and dynamic provisioned ERP systems. In this demonstration we want to show SLA management for hosted ERP systems from three perspectives; the customer, the sales, and IT administrator perspective. Aspects of the negotiation, planning and provisioning workflow, which involves all stakeholders, are outlined.".
- 5 abstract "Traditional approaches for semantic storage and analysis are facing their limits on the handling of enormous data amounts of today’s applications. We believe that a more radical departure from contemporary architectures of stores is necessary to satisfy that central scalability requirement. One of the most promising new schools of thought in system design are swarm intelligent and swarm-based approaches for data distribution and organization. In this paper, we describe our current work on a swarm-based storage service for Semantic Web data.".
- 6 abstract "Maximizing network throughput via back-pressure routing is the subject of a considerable body of literature. However, the associated optimal throughput guarantees come at the cost of excessively high delays. Though there exists some proposals in literature for improving the delay performance, the impact of these proposals on the delay performance is very little in loopy networks (i.e., networks with routing loops). In this demonstration, with series of experiments over different topologies, we show that the basic back-pressure algorithm and two of its variants for improving delay performance induce extensive routing loops with associated high packet delay even in simple network topologies.".
- 8 abstract "One of the current barriers towards realizing the huge potential of Future Internet is the protection of sensitive information, i.e., the ability to selectively expose (or hide) information to (from) users depending on their access privileges. In this work we discuss the experiments conducted with our repository independent, portable across platforms system that supports fine-grained enforcement of RDF access control.".
- 9 abstract "This submission will provide interested parties with the opportunity to learn more about the FP7 PERIMETER project. Details of the testing and experimental methodologies and the role of testbed activities involved in this Quality of Experience network mobility project will be given. A demonstration and explanation of the Quality of Experience management system of PERIMETER will be provided and will allow visitors to interact with the system using a Google Android device with the PERIMETER middleware installed. This will be illustrated with the aid of the purpose built PERIMETER Visualisation Tool.".
- 1 abstract "One of the current barriers towards realizing the huge potential of Future Internet is the protection of sensitive information, i.e., the ability to selectively expose (or hide) information to (from) users depending on their access privileges. Given that RDF has established itself as the de facto standard for data representation over the Web, our work focuses on controlling access to RDF data. We present a high-level access control specification language that allows fine-grained specification of access control permissions (at triple level) and formally define its semantics. We adopt an annotation-based enforcement model, where a user can explicitly associate data items with annotations specifying whether the item is accessible or not. In addition, we discuss the implementation of our framework, propose a set of dimensions that should be considered when defining a benchmark to evaluate the different access control enforcement models and present the results of our experiments conducted on different Semantic Web platforms.".
- 2 abstract "Imagine what kind of applications become possible when our mobile devices not only present data but provide valuable information to other users. Users become able to instantaneously create services and to publish content and knowledge on their own mobile device, which can be discovered and accessed remotely by other mobile users in a simple way. To achieve the vision of customizable and context aware user-generated mobile services, we present a mobile authoring tool for end-users to create, customize and deploy mobile services while on-the-go. This tool is designed to allow users with different levels of technical expertise to create mobile services. The paper also gives insight on the performed usability evaluations, namely user interviews and an online survey.".
- 3 abstract "Given the significant amount of personal information available on the Web, verifying its correct use emerges as an important issue. When personal information is published, it should be later used under a set of usage policies. If these policies are not followed, sensitive data could be exposed and used against its owner. Under these circumstances, processing transparency is desirable since it allows users to decide whether information is used appropriately. It has been argued that data provenance can be used as the mechanism to underpin such a transparency. Thereby, if provenance of data is available, processing becomes transparent since the provenance of data can be analysed against usage policies to decide whether processing was performed in compliance with such policies. The aim of this paper is to present a Provenance-based Compliance Framework that uses provenance to verify the compliance of processing to predefined information usage policies. It consists of a provenance-based view of past processing of information, a representation of processing policies and a comparison stage in which the past processing is analysed against the processing policies. This paper also presents an implementation using a very common on-line activity: on-line shopping.".
- 1 abstract "The Semantic Web (SW) is still far from realising its full potential,partly because it is still lacking enough high quality SW representations of information.We argue that a step in the right direction is fostering people’s capability to fluently create high quality SW representations of the information they generate during problem solving processes. To foster such a capability, we propose a game in which teams compete in creating the best translations of texts into SW representations. Although playing the game is in itself already a way to foster such a capability, we moreover pursue learning from the game which are the most successful translation strategies (embodied by “constitutions”) so that they can also be used by people outside a game setting.".
- 2 abstract "The importance of incentives and socially based motivation for metadata generation should not distract attention entirely from the need to design tools for metadata generation that use every means available to maximize the efficiency and intrinsic motivation of the individual annotator. The popular application domain of (individual) photo management has recently given rise to a number of strategies and methods that can serve as a source of inspiration for the design of metadata generation support for the semantic web. This position paper offers a brief synthesis of relevant work that is intended to serve as a basis for the representation of this perspective at the Insemtive 2008 workshop.".
- 3 abstract "Many use cases for the SemanticWeb assume the availability of public metadata. However, research has not yet addressed in a satisfactory manner why and how metadata is published on the Semantic Web. We analyze several reasons and barriers for creating and sharing semantic metadata. In particular, we address the issue of how metadata from private spaces can diffuse into the public Semantic Web. Therefore we introduce the concept of inverse semantic search – an approach which aggregates information needs to motivate information providers to share private metadata.".
- 4 abstract "This paper argues for collaborative incremental augmentation of text retrieval as an approach that can be used to immediately show the benefits of relatively heavyweight knowledge formalization in the context of Web 2.0 style collaborative knowledge formalization. Such an approach helps to overcome the “Curse of Prepayment”; i.e. the hitherto necessary very large initial investment in formalization tasks before any benefit of Semantic Web technologies is visible. Some initial ideas about the architecture of such a system are presented and it is placed within the overall emerging trend of “people powered search”.".
- 5 abstract "The Swedish Oral Medicine Web (SOMWeb) is an online system built to support knowledge sharing between oral medicine practitioners who hold monthly telephone conferences to discuss difficult and interesting cases. Semantic Web technologies are used to model the templates used for case entry, the ontology of values used in filling in cases, and community data. To study the practitioners’ use and perceptions of the collaboration and the SOMWeb system, we have used observations of teleconferences, interviews with participants, and an online questionnaire.These are analyzed to provide a picture of the participants’ opinions about the structured case entry and why they contribute or do not contribute cases. This is followed by a discussion on future work where the value ontology is made available for community editing and structuring, and incentives for user contributions to this process.".
- 6 abstract "Folksonomy is a method of classifying content, and it iswidely used in some web services. It allows users to choose tags (keywords or terms assigned to specific content) freely and to search content by referring to the tags. Compared to existing classification methods, folksonomy reflects the users’ intention more directly because of its unlimited vocabulary and multiple tags for one content item. Moreover, it has a useful characteristic where tags represent the description of the content. Although tags are intended to be a rich semantic description of web content, machines cannot understand what the tags mean because they are just keywords. We describe a method to extract the concept related to the tag in a machine-understandable way by focusing on the features of content annotated with each tag. In particular, we target the problem of extracting the temporal and spatial concepts of the tags on Flickr, a popular photo sharing service by looking at the date and location distributions of photos for each tag. We evaluated the concept extracting method on a snapshot of actual Flickr data and show that it can identify a tags’ concept in a manner similar to the way a person can.".
- 905 abstract "Many enterprise systems face the same kind of information processing problems that exist on the web in general, and creating semantic solutions often involve constructing an enterprise ontology. Ontology engineering in turn needs to be semi-automatic in order to reduce the effort and need for expertise. By introducing knowledge reuse in ontology construction the construction effort can be further reduced and the quality of the ontology output increased. The proposed research focuses on a hybrid approach for ontology construction based on the methodology of case-based reasoning in combination with ontology patterns.".
- 910 abstract "This paper describes a system capable of semi-automatically filling an XML template from free texts in the clinical domain (practice guidelines). The XML template includes semantic information not explicitly encoded in the text (pairs of conditions and actions/recommendations). Therefore, there is a need to compute the exact scope of conditions over text sequences expressing the required actions. We present a system developed for this task. We show that it yields good performance when applied to the analysis of French practice guidelines.".
- 915 abstract "The move towards a semantic web has been in progress for many years and more recently there have been applications that make use of semantic web technology. One of the features that made the Web so easy to use is the ability to search web pages in a matter of seconds through the use of search engines. Now that the use of OWL and RDF as a knowledge representation format is increasing, the possibility appears to improve the quality of searching by using the semantic web to enhance the 'ordinary' Web. This paper outlines an architecture for using distributed knowledge bases to assist and improve searching on the web.".
- 920 abstract "Recent work in the field of middleware technology proposes semantic spaces as a tool for coping with the scalability, heterogeneity and dynamism issues of large scale distributed environments. Reflective middleware moreover offers answers to the needs for adaptivity and self-determination of systems. Based on experiences with traditional middleware we argue that ontology-driven management is a major advancement for semantic spaces and provides the fundamental means for reflection. In this research we look at ontology-based metadata to turn semantic spaces into reflective middleware.".
- 925 abstract "The World Wide Web (WWW) now is widely used as a universal medium for information exchange. Semantic interoperability among different information systems in the WWW is limited due to information heterogeneity problem. Ontologies have been suggested as a way to solve this problem. Given that no universal ontology exists for the WWW, work has focused on finding semantic correspondences between similar elements of different ontologies, i.e., ontology mapping. Ontology mapping can be done either by hand or using automated tools. Manual mapping becomes impractical as the size and complexity of ontologies increases. Fully or semi-automated mapping approaches have been examined by different research studies, e.g., analyzing linguistic information of elements in ontologies, treating ontologies as structural graphs, applying heuristic rules to look for specific mapping patterns and machine learning techniques. This paper proposes a new generic and scalable ontology mapping approach, which takes advantage of propagation theory, information retrieval technique and artificial intelligence model to solve ontology mapping problem. The major procedures of the proposed approach are profile enrichment, profile propagation, IR-based profile mapping and the interactive activation network based constraints satisfaction. The profile enrichment builds a profile for each element in the ontologies, and thus enriches its information. The profile propagation exploits the neighboring information of each element by passing the profile of the ancestors, descendants or siblings of the element to the profile of itself, with different weights. The IR-based profile mapping process compares the cosine similarity of elements’ profiles in a vector space model. Finally the interactive activation network is integrated to deal with the constraints satisfaction problems in the context of ontology mapping.".
- 930 abstract "This paper proposes research on a probabilistic framework for information integration and retrieval in the Semantic Web. The framework aims at integrating probabilistic and deterministic DL and LP knowledge bases.".
- 935 abstract "We propose a framework to model negotiations among agents in scenarios withfully incomplete information. We address multi-issue bilateral negotiations, where issues are expressed and related to each other via an OWL ontology. In particular we use OWL DL sub-language - as it is the "official language" of the Semantic Web communities - and exploit its formal semantics based on the Description Logic SHOIN(D). Agents goals are expressed through OWL-DL as (complex) concept expressions, and the worth of goals as weights over concepts. We adopt a very general setting with incomplete information by letting agents keep both goals and worths of goals as private information. Moreover we propose a framework through which agents can negotiate and reach an agreement. Agents, according to the protocol, can pursue different sets of strategies, depending on their type.".
- 940 abstract "Our research explores the possibilities for factoring culture into user models, working towards cultural adaptivity in the Semantic Web. The aim is to automate the localization process by acquiring details about an individual’s cultural background. The paper shows how we have based our approach on research findings from related fields, outlining the connection between the cultural user and domain model ontologies, and a test environment. Furthermore, we describe a preliminary plan of how to validate our cultural user model by comparison to our previous experience in the manual adaptation of imbuto.".
- 945 abstract "Only few well-maintained domain ontologies can be found on the Web. The reasons for the lack of useful domain ontologies include that (1) informal means to convey intended meaning more efficiently are used for ontology specification only to a very limited extent, (2) many relevant domains of discourse show a high degree of conceptual dynamics, (3) ontology representation languages are hard to understand for the majority of (potential) ontology users, and (4) the community does not have control over the ontology evolution. In my thesis, on which I have been working for 9 months, I propose to (1) base a methodology for community-grounded ontology building on the culture and philosophy of Wikis by giving users who have no or little expertise in ontology engineering the opportunity to contribute in all stages of the ontology lifecycle and (2) exploit the combination of human and computational intelligence to discover and resolve inconsistencies and align lightweight domain ontologies. The contribution of my thesis is a methodology for community-grounded building and evolution of lightweight domain ontologies.".
- 950 abstract "Unsatisfactory performance of ontology mapping methods on many tasks seems to indicate the need for employing some background knowledge. My PhD work focuses on exploiting pattern-like structures for this purpose. Design patterns are related to different styles of modelling ontology structure and/or naming of concepts. On the other hand, mapping patterns emerge as result of the mapping process. Preliminary results and schedule of future work are presented.".
- 955 abstract "This paper gives an outline of my PhD thesis which describes the integration of managing uncertainty into ontology mapping. Ontology mapping is one of the most important tasks for ontology interoperability and its main aim is to find semantic relationships between entities (i.e. concept, attribute, and relation) of two ontologies, However, in the process of mapping, uncertainty and incompleteness of semantics in the syntactic representation and description of relations between entities in ontologies will lead to imprecise results. If we want to obtain better results, it becomes more significant for the ontology mapping to be able to deal with uncertainty.".
- 960 abstract "The use of tags to describe Web resources in a collaborative manner has experienced rising popularity among Web users in recent years. The product of such activity is given the name folksonomy, which can be considered as a scheme of organizing information in the users' own way. This research work attempts to analyze tripartite graphs - graphs involving users, tags and resources - of folksonomies and discuss how these elements acquire their semantics through their associations with other elements, a process we call mutual contextualization. By studying such process, we try to identify solutions to problems such as tag disambiguation, retrieving documents of similar topics and discovering communities of users. This paper describes the basis of the research work, mentions work done so far and outlines future plans.".
- 701 abstract "Governments often hold very rich data and whilst much of this information is published and available for re-use by others, it is often trapped by poor data structures, locked up in legacy data formats or in fragmented databases. One of the great benefits that Semantic Web (SW) technology offers is facilitating the large scale integration and sharing of distributed data sources. At the heart of information policy in the UK, the Office of Public Sector Information (OPSI) is the part of the UK government charged with enabling the greater re-use of public sector information. This paper describes the actions, findings, and lessons learnt from a pilot study, involving several parts of government and the public sector. The aim was to show to government how they can adopt SW technology for the dissemination, sharing and use of its data.".
- 715 abstract "DBpedia is a community effort to extract structured information from Wikipedia and to make this information available on the Web. DBpedia allows you to ask sophisticated queries against datasets derived from Wikipedia and to link other datasets on the Web to Wikipedia data. We describe the extraction of the DBpedia datasets, and how the resulting information can be made available on the Web for humans and machines. We describe some emerging applications from the DBpedia community and show how website operators can reduce costs by facilitating royalty-free DBpedia content within their sites. Finally, we present the current status of interlinking DBpedia with other open datasets on the Web and outline how DBpedia could serve as a nucleus for an emerging Web of open data sources.".
- 729 abstract "This paper presents a semantic case-based reasoning framework for text categorization. Text categorization is the task of classifying text documents under predened categories. Accidentology is our application eld and the goal of our framework is to classify documents describing real road accidents under predened road accident prototpypes, which also are described by text documents. Accidents are described by accident reports while accident prototypes are described by accident scenarios. Thus, text categorization is done by assigning each accident report to an accident scenario, which highlights particular mechanisms leading to accident. We propose a textual case based reasoning approach (TCBR), which allows us to integrate both textual and domain knowledge aspects inorder to carry out this categorization. CBR solves a new problem (target case) by identifying its similarity to one or several previously solved problems (source cases) stored in a case base and by adapting their known solutions. Cases of our framework are created from text. Most of TCBR applications create cases from text by using Information Retrieval techniques, which leads to knowledge-poor descriptions of cases. We show that using semantic resources (two ontology of accidentology) makes possible to overcome this diculty, and allows us to enrich cases by using formal knowledge. In this paper, we argue that semantic resources are likely to improve the quality of cases created from text, and, therefore, such resources can support the reasoning cycle. We illustrate this claim with our framework developed to classify documents in the accidentology domain.".
- 743 abstract "We describe a novel approach by which software can assess the ability of a confederation of heterogeneous systems to interoperate to achieve a given purpose. The approach uses ontologies and knowledge bases (KBs) to capture the salient characteristics of systems, on the one hand, and of tasks for which these systems will be employed, on the other. Rules are used to represent the conditions under whichthe capabilities provided by systems can fulfill the capabilities needed to support the roles and interactions that make up each task. An Analyzer component employs these KBs and rules to determine if a given confederation will be adequate, to generate suitable confederations from a collection of available systems, to pre-diagnose potential interoperability problems that might arise, and to suggest system configuration options that will help to make interoperability possible. We have demonstrated the feasibility of this approach using a prototype Analyzer and KBs.".
- 757 abstract "This paper introduces a framework to add a semantic web layer to legacy organizational information, and describes its application to the use case provided by the Italian National Research Council (CNR) intraweb. Building on a traditional web-based view of information from different legacy databases, we have performed a semantic porting of data into a knowledge base, dependent on an OWL domain ontology. We have enriched the knowledge base by means of text mining techniques, in order to discover on-topic relations. Several reasoning techniques have been applied, in order to infer relevant implicit relationships. Finally, the ontology and the knowledge base have been deployed on a semantic wiki by means of the WikiFactory tool, which allows users to browse the ontology and the knowledge base, to introduce new relations, to revise wrong assertions in a collaborative way, and to perform semantic queries. In our experiments, we have been able to easily implement several functionalities, such as expert finding, by simply formulating ad-hoc queries from either an ontology editor or the semantic wiki interface. The result is an intelligent and collaborative front end, which allow users to add information, fill gaps, or revise existing information on a semantic basis, while keeping the knowledge base automatically updated.".
- 771 abstract "This paper shows how semantic web techniques can be applied to solving problems of distributed content creation, discovery, linking, aggregation, and reuse in health information portals, both from end-user’s and content publishers’s viewpoints. As a case study, the national semantic health portal HEALTHFINLAND is presented. It provides citizens with intelligent searching and browsing services to reliable and up-to-date health information created by various health organizations in Finland. The system is based on a shared semantic metadata schema, ontologies, and mash-up ontology services. The content includes metadata of thousands of web documents such as web pages, articles, reports, campaign information, news, services, and other information related to health.".
- 785 abstract "As an increasing number of applications on the web contain some elements of spatial data, there is a need to efficiently integrate Semantic Web technologies and spatial data processing. This paper describes a prototype system for storing spatial data and Semantic Web data together in a SPatially-AUgmented Knowledgebase (SPAUK) without sacrificing query efficiency. The goals are motivated through use several use cases. The prototype’s design and architec-ture are described, and resulting performance improvements are discussed.".
- 795 abstract "Semantic Web conferences such as ESWC and ISWC offer prime opportunities to test and showcase semantic technologies. Conference metadata about people, papers and talks is diverse in nature and neither too small to be uninteresting or too big to be unmanageable. Many metadata-related challenges that may arise in the Semantic Web at large are also present here. Metadata must be generated from sources which are often unstructured and hard to process, and may originate from many different players, therefore suitable workflows must be established. Moreover, the generated metadata must use appropriate formats and vocabularies, and be served in a way that is consistent with the principles of linked data. This paper reports on the metadata efforts from ESWC and ISWC, identifies specific issues and barriers encountered during the projects, and discusses how these were approached. Recommendations are made as to how these may be addressed in the future, and we discuss how these solutions may generalize to metadata production for the Semantic Web at large.".
- 809 abstract "This paper describes a large case study that explores the applicability of ontology reasoning to problems in the medical domain. We investigate whether it is possible to use such reasoning to automate common clinical tasks that are currently labor intensive and error prone, and focus our case study on improving cohort selection for clinical trials. An obstacle to automating such clinical tasks is the need to bridge the semantic gulf between raw patient data, such as laboratory tests or specific medications, and the way a clinician interprets this data. Our key insight is that matching patients to clinical trials can be formulated as a problem of semantic retrieval. We describe the technical challenges to building a realistic case study, which include problems related to scalability, the integration of large ontologies, and dealing with noisy, inconsistent data. Our solution is based on the SNOMED CT ontology, and scales to one year of patient records (approx. 240,000 patients).".
- 823 abstract "An ontology provides a precise specification of the vocabulary used by a community of interest (COI). Multiple communities of interest may describe the same concept using the same or different terms. When such communities interact, ontology alignment and translation is required. This is typically a time consuming process. This paper describes Snoggle, an open source tool designed to ease development of ontology translation rules, and discusses its application to geospatial ontologies.".
- 837 abstract "Business Intelligence (BI) requires the acquisition and aggregation of key pieces of knowledge from multiple sources in order to provide valuable information to customers or feed statistical BI models and tools. The massive amount of information available to business analysts makes information extraction and other natural language processing tools key enablers for the acquisition and use of that semantic information. We describe the application of ontology-based extraction and merging in the context of a practical e-business application for the EU MUSING Project where the goal is to gather international company intelligence and country/region information. The results of our experiments so far are very promising and we are now in the process of bulding a complete end-to-end solution.".
- 851 abstract "Data warehouse is now widely used in business analysis and decision making processes. To adapt the rapidly changing business environment, we develop a tool to make data warehouses more business-friendly by using Semantic Web technologies. The main idea is to make business semantics explicit by uniformly representing the business metadata (i.e. conceptual enterprise data model and multidimensional model) with an extended OWL language. Then a mapping from the business metadata to the schema of the data warehouse is built. When an analysis request is raised, a customized data mart with data populated from the data warehouse can be automatically generated with the help of this built-in knowledge. This tool, called Enterprise Information Asset Workbench (EIAW), is deployed at the Taikang Life Insurance Company, one of the top five insurance companies of China. User feedback shows that OWL provides an excellent basis for the representation of business semantics in data warehouse, but many necessary extensions are also needed in the real application. The user also deemed this tool very helpful because of its flexibility and speeding up data mart deployment in face of business changes.".