004 Datenverarbeitung; Informatik
Filtern
Erscheinungsjahr
Dokumenttyp
- Ausgabe (Heft) zu einer Zeitschrift (36)
- Dissertation (15)
- Diplomarbeit (4)
- Masterarbeit (4)
- Studienarbeit (4)
- Bachelorarbeit (1)
Sprache
- Englisch (64) (entfernen)
Schlagworte
- Bluetooth (4)
- Software Engineering (4)
- Knowledge Compilation (3)
- Campus Information System (2)
- E-KRHyper (2)
- Equality (2)
- Modellgetriebene Entwicklung (2)
- Ontology (2)
- Petri-Netze (2)
- Semantic Web (2)
- Theorem Proving (2)
- University (2)
- constraint logic programming (2)
- hybrid automata (2)
- probability propagation nets (2)
- API (1)
- API analysis (1)
- API-Analyse (1)
- AUTOSAR (1)
- Abduktion <Logik> (1)
- Adaptation (1)
- Algorithm Engineering (1)
- Anpassung (1)
- Architektur <Informatik> (1)
- Auditing (1)
- Augmented Reality (1)
- Automated Theorem Proving (1)
- Automated Theorem Proving Systems (1)
- Automotive Systems (1)
- Bayes Procedures (1)
- Befahrbarkeit (1)
- Border Gateway Protocol (1)
- Border Gateway Protocol 4 (1)
- Calculus (1)
- Communication Networks (1)
- Computer Security (1)
- Computersicherheit (1)
- Computersimulation (1)
- Conference (1)
- Context-aware processes (1)
- DPLL procedure (1)
- Data Mining (1)
- Deduktion (1)
- Description Logics (1)
- Destiny (1)
- Developer profiling (1)
- Dijkstras Algorithmus (1)
- Dimension 3 (1)
- Diskrete Simulation (1)
- Distributed Environments (1)
- Distributed process execution (1)
- Driver Assistance Systems (1)
- Empirical Studies (1)
- Emulation (1)
- Enhanced Reality (1)
- Entwickler Profil (1)
- Fahrzeug (1)
- Formal Methods (1)
- Formale Methoden (1)
- Formale Ontologie (1)
- Fragebeantwortung (1)
- Gelände (1)
- Grounded Theory (1)
- Hindernis (1)
- Horn Clauses (1)
- Human-Computer Interaction (1)
- Hyper Tableau Calculus (1)
- IASON (1)
- IT Guru (1)
- Information Retrieval (1)
- Intelligent Information Network (1)
- Interactive Video Retrieval (1)
- JML (1)
- Java (1)
- Java Modeling Language (1)
- Java. Programmiersprache (1)
- KRHyper (1)
- Klassifikation (1)
- Knowledge (1)
- Laser (1)
- Logischer Schluss (1)
- MIA (1)
- MPEG-7 (1)
- Mensch-Maschine-Interaktion (1)
- Mining (1)
- Mobile Information Systems (1)
- Multi-robot System (1)
- Multiagent System (1)
- Multiagentensysteme (1)
- Netzwerk (1)
- Netzwerkanalyse (1)
- Neuronales Netz (1)
- OPNET (1)
- OWL <Informatik> (1)
- Ontologie <Wissensverarbeitung> (1)
- Oracle Generation (1)
- Oraklegenerierung (1)
- Personalised Information Systems (1)
- Petri Nets (1)
- Petri net (1)
- Petrinetz (1)
- Policy Language (1)
- Probability (1)
- Probability propagation nets (1)
- Process tracing (1)
- Propagation (1)
- Provenance (1)
- Prädikatenlogik (1)
- Query Expansion (1)
- RDF Graphs (1)
- RDF modeling (1)
- Relevance Feedback (1)
- Resource Description Framework (RDF) (1)
- Robocup 2008 (1)
- Roboter (1)
- Routing Information Protocol (RIP) (1)
- Routing Loops (1)
- Routing with Metric based Topology Investigation (RMTI) (1)
- SOA (1)
- SPARQL (1)
- Schlussfolgern (1)
- Semantics (1)
- Serviceorientierte Architektur (1)
- Software Language (1)
- Software Technology (1)
- Softwarearchitektur (1)
- Softwaretest (1)
- Softwaretesting (1)
- Specification (1)
- Spezifikation (1)
- Stochastic Logic (1)
- Straßenzustand (1)
- TAP (1)
- TRECVID (1)
- Tableau Calculus (1)
- Technologischer Raum (1)
- Test Generation (1)
- Testgenerierung (1)
- Theorem prover (1)
- Theorembeweiser (1)
- Tokens (1)
- UML (1)
- Unified Modeling Language (UML ) (1)
- Vegetation distribution (1)
- Verifikation (1)
- Wahrscheinlichkeitsrechnung (1)
- Web Ontology Language (OWL) (1)
- Werbung (1)
- Wissensmanagement (1)
- World Wide Web 2.0 (1)
- XML (1)
- automated theorem prover (1)
- backpropagation (1)
- bias (1)
- classification (1)
- concurrency (1)
- data mining (1)
- deductive (1)
- e-learning (1)
- event model (1)
- event-based systems (1)
- first-order logic (1)
- folksonomies (1)
- gaze information (1)
- grassland (1)
- hybrid systems (1)
- hybride Automaten (1)
- iCity project (1)
- image semantics (1)
- information retrieval (1)
- knowledge management system (1)
- landmarks (1)
- living book (1)
- model generation (1)
- multi-agent systems (1)
- multiagent systems (1)
- networks (1)
- neural (1)
- ontology (1)
- personal information management (1)
- persönliches Informationsmanagement (1)
- privacy protection (1)
- public key infrastructure (1)
- question answering (1)
- recurrent (1)
- semantic desktop (1)
- semantics (1)
- semantischer Desktop (1)
- sequent calculi (1)
- simulation (1)
- summative evaluation (1)
- tag recommendation (1)
- tagging (1)
- vegetation modeling (1)
- verification (1)
- web 2.0 (1)
Institut
- Institut für Informatik (64) (entfernen)
Querying for meta knowledge
(2008)
The Semantic Web is based on accessing and reusing RDF data from many different sources, which one may assign different levels of authority and credibility. Existing Semantic Web query languages, like SPARQL, have targeted the retrieval, combination and reuse of facts, but have so far ignored all aspects of meta knowledge, such as origins, authorship, recency or certainty of data, to name but a few. In this paper, we present an original, generic, formalized and implemented approach for managing many dimensions of meta knowledge, like source, authorship, certainty and others. The approach re-uses existing RDF modeling possibilities in order to represent meta knowledge. Then, it extends SPARQL query processing in such a way that given a SPARQL query for data, one may request meta knowledge without modifying the query proper. Thus, our approach achieves highly flexible and automatically coordinated querying for data and meta knowledge, while completely separating the two areas of concern.
Geographisches Cluster-basiertes Routing ist ein aktueller Ansatz, wenn es um das Entwicklen von effizienten Routingalgorithmen für drahtlose ad-hoc Netzwerke geht. Es gibt bereits eine Anzahl an Algorithmen, die Nachrichten nur auf Basis von Positionsinformationen durch ein drahtloses ad-hoc Netzwerk routen können. Darunter befinden sich sowohl Algorithmen, die auf das klassische Beaconing setzen, als auch Algorithmen, die beaconlos arbeiten (keine Informationen über die Umgebung werden benötigt, außer der eigenen Position und der Position des Ziels). Geographisches Routing mit Auslieferungsgarantie kann auch auf Overlay-Graphen durchgeführt werden. Bisher werden die dafür benötigten Overlay-Graphen nicht reaktiv konstruiert.
In dieser Arbeit wird ein reaktiver Algorithmus, der Beaconless Cluster Based Planarization Algorithmus (BCBP), für die Konstruktion eines planaren Overlay-Graphen vorgestellt, der die benötigte Anzahl an Nachrichten für die Konstruktion eines planaren Overlay-Graphen, und demzufolge auch Cluster-basiertes geographishes Routing, deutlich reduziert. Basierend auf einem Algorithmus für Cluster-basierte Planarisierung, konstruiert er beaconlos einen planaren Overlay-Graphen in einem unit disk graph (UDG). Ein UDG ist ein Modell für ein drahtloses Netzwerk, bei dem alle Teilnehmer den gleichen Senderadius haben.
Die Evaluierung des Algorithmus zeigt, dass er wesentlich effizienter ist als die Baecon-basierte Variante. Ein weiteres Ergebnis dieser Arbeit ist ein weiterer beaconloser Algorithmus (Beaconless LLRAP (BLLRAP)), für\r\nden zwar die Planarität, aber nicht die Konnektivität nachgewiesen werden konnte.
Die Diffusionsbildgebung misst die Bewegung von Wassermolekülen in Gewebe mittelsrnvariierender Gradientenfelder unter Verwendung der Magnetresonanztomographie(MRT). Diese Aufnahmetechnik stellt eine große Chance für in vivo Untersuchung von neuronalen Bahnen dar, da das lokale Diffusionsprofil Rückschlüsse über die Position und Richtung von Nervenbahnen erlaubt. Zu den Anwendungsgebieten der Diffusionsbildgebung zählt die Grundlagenforschung in den Neurowissenschaften, in denen Nervenbahnen als Verbindungen kortikaler Areale bestimmt werden, und die neurochirurgische Operationsplanung, in der rekonstruierte Bahnen als Risikostrukturen für Interventionen angesehen werden.
Die Diffusionstensor-MRT (DT-MRT) ist aufgrund ihrer schnellen Aufnahme- und Rekonstruktionsgeschwindigkeit derzeitig klinischer Standard zur Bestimmung von Nervenbahnen. Jedoch erlaubt die DT-MRT nicht die Darstellung von komplexen intravoxel Diffusionsverteilungen. Daher etablierte sich eine weitere Modellierungstechnik, die als High Angular Resolution Diffusion Imaging (HARDI) bekannt ist. HARDITechniken erhielten wachsendes Interesse in den Neurowissenschaften, da sie großes Potential zur exakteren Darstellung der Nervenbahnen im menschlichen Gehirn besitzen.
Um die Vorteile von HARDI-Techniken gegenüber DT-MRT voll auszuschöpfen, werden fortgeschrittene Methoden zur Rekonstruktion und Visualisierung der Bahnen benötigt. In der vorliegenden Arbeit werden neue Techniken vorgestellt, welche zur aktuellen Forschung hinsichtlich der Verarbeitung und Visualisierung von Diffusionsbildgebungsdaten beitragen. Ansätze zur Klassifizierung, Traktographie und Visualisierung wurden entwickelt um eine aussagekräftige Exploration neuronaler Bahnen und deren Beschaffenheit zu ermöglichen. Des Weiteren wurde eine interaktive Software für die neurochirurgische Operationsplanung implementiert, welche Nervenbahnen als Risikostrukturen berücksichtigt.
Die vorgestellten Forschungsergebnisse bieten einen erweiterten und aufgabenorientierten Einblick in neuronale Verbindungen sowohl für Neurowissenschaftler als auch für Neurochirurgen und tragen zum Einsatz von HARDI-Techniken in einer klinischen Umgebung bei.
The University of Koblenz-Landau would like to apply for participation in the RoboCup Mixed Reality League in Suzhou, China 2008. Our team is composed of ten team members and two supervisors. All members are graduate students of Computational Visualistics. Our supervisors are Ph.D. candidates currently researching in the working groups of artificial intelligence and computer graphics.
Die hohen Infrastrukturkosten machen das Überprüfen von Theorien bezüglich großer Rechnernetze zu einer schwierigen und teuren Aufgabe. Ein möglicher Ansatz dieses Problem zu beheben ist die Verwendung von virtueller anstelle von physikalischer Infrastrukur. OPNets IT Guru ist ein Programm, das entworfen wurde zur Simulation großer Netze und zur Repräsentation relevanter Informationen. Es gestattet großflächige Änderungen zu testen oder Theorien zu überpruefen ohne den Aufwand einer physikalischen Infrastruktur.
The E-KRHyper system is a model generator and theorem prover for first-order logic with equality. It implements the new E-hyper tableau calculus, which integrates a superposition-based handling of equality into the hyper tableau calculus. E-KRHyper extends our previous KRHyper system, which has been used in a number of applications in the field of knowledge representation. In contrast to most first order theorem provers, it supports features important for such applications, for example queries with predicate extensions as answers, handling of large sets of uniformly structured input facts, arithmetic evaluation and stratified negation as failure. It is our goal to extend the range of application possibilities of KRHyper by adding equality reasoning.
Generalized methods for automated theorem proving can be used to compute formula transformations such as projection elimination and knowledge compilation. We present a framework based on clausal tableaux suited for such tasks. These tableaux are characterized independently of particular construction methods, but important features of empirically successful methods are taken into account, especially dependency directed backjumping and branch local operation. As an instance of that framework an adaption of DPLL is described. We show that knowledge compilation methods can be essentially improved by weaving projection elimination partially into the compilation phase.
This volume contains those research papers presented at the Second International Conference on Tests and Proofs (TAP 2008) that were not included in the main conference proceedings. TAP was the second conference devoted to the convergence of proofs and tests. It combines ideas from both areas for the advancement of software quality. To prove the correctness of a program is to demonstrate, through impeccable mathematical techniques, that it has no bugs; to test a program is to run it with the expectation of discovering bugs. On the surface, the two techniques seem contradictory: if you have proved your program, it is fruitless to comb it for bugs; and if you are testing it, that is surely a sign that you have given up on any hope of proving its correctness. Accordingly, proofs and tests have, since the onset of software engineering research, been pursued by distinct communities using rather different techniques and tools. And yet the development of both approaches leads to the discovery of common issues and to the realization that each may need the other. The emergence of model checking has been one of the first signs that contradiction may yield to complementarity, but in the past few years an increasing number of research efforts have encountered the need for combining proofs and tests, dropping earlier dogmatic views of their incompatibility and taking instead the best of what each of these software engineering domains has to offer. The first TAP conference (held at ETH Zurich in February 2007) was an attempt to provide a forum for the cross-fertilization of ideas and approaches from the testing and proving communities. For the 2008 edition we found the Monash University Prato Centre near Florence to be an ideal place providing a stimulating environment. We wish to sincerely thank all the authors who submitted their work for consideration. And we would like to thank the Program Committee members as well as additional referees for their great effort and professional work in the review and selection process. Their names are listed on the following pages. In addition to the contributed papers, the program included three excellent keynote talks. We are grateful to Michael Hennell (LDRA Ltd., Cheshire, UK), Orna Kupferman (Hebrew University, Israel), and Elaine Weyuker (AT&T Labs Inc., USA) for accepting the invitation to address the conference. Two very interesting tutorials were part of TAP 2008: "Parameterized Unit Testing with Pex" (J. de Halleux, N. Tillmann) and "Integrating Verification and Testing of Object-Oriented Software" (C. Engel, C. Gladisch, V. Klebanov, and P. Rümmer). We would like to express our thanks to the tutorial presenters for their contribution. It was a team effort that made the conference so successful. We are grateful to the Conference Chair and the Steering Committee members for their support. And we particularly thank Christoph Gladisch, Beate Körner, and Philipp Rümmer for their hard work and help in making the conference a success. In addition, we gratefully acknowledge the generous support of Microsoft Research Redmond, who financed an invited speaker.
The model evolution calculus
(2004)
The DPLL procedure is the basis of some of the most successful propositional satisfiability solvers to date. Although originally devised as a proof procedure for first-order logic, it has been used almost exclusively for propositional logic so far because of its highly inefficient treatment of quantifiers, based on instantiation into ground formulas. The recent FDPLL calculus by Baumgartner was the first successful attempt to lift the procedure to the first-order level without resorting to ground instantiations. FDPLL lifts to the first-order case the core of the DPLL procedure, the splitting rule, but ignores other aspects of the procedure that, although not necessary for completeness, are crucial for its effectiveness in practice. In this paper, we present a new calculus loosely based on FDPLL that lifts these aspects as well. In addition to being a more faithful litfing of the DPLL procedure, the new calculus contains a more systematic treatment of universal literals, one of FDPLL's optimizations, and so has the potential of leading to much faster implementations.
Dualizing marked Petri nets results in tokens for transitions (t-tokens). A marked transition can strictly not be enabled, even if there are sufficient "enabling" tokens (p-tokens) on its input places. On the other hand, t-tokens can be moved by the firing of places. This permits flows of t-tokens which describe sequences of non-events. Their benefiit to simulation is the possibility to model (and observe) causes and effects of non-events, e.g. if something is broken down.