GraphDB's Visual Graph does not show any visible connections - graphdb

Here's what I have (using a Mac):
An ontology was created in Protege (v5), incorporating some classes
from established ontologies.
All classes were given the same Base URI.
HermiT, FacT++, Pellet, and Ontop plugin reasoners were used to test the ontology in Protege, and no errors or messages were returned.
Protege's Ontop plugin (v1.8) was used to map the classes to
postgres RDBS column headers (table_column) and exported the R2RML
mapping file.
The Ontop CLI was used to materialize triples through a
jdbc postgres connection the ontology in turtle format, according to
the R2RML .ttl mappings file.
Here's where my interactions with GraphDB (v8.2) start:
The GraphDB repository was created with an OWL2-RL profile with the 'read only' unchecked.
All ontology and materialized data triples were uploaded to GraphDB through graphdb-import folder in turtle format. All options checked on upload, and the same Base URI was provided.
The Class Hierarchy diagram suggests that the imported triples are properly structured.
The Class Relationship circular network graph shows reasonable
relationships between the largest 10 classes.
The Domain-Range graph shows a number of in-going and out-going connections to nodes/individuals with predicates.
SPARQL queries appear to return expected values.
Here's my problem: When I enter the Visual Graph, I don't see anything. When I search and select a value that should have nodes, all I see is the single node with a message "This node has no visible connections."
What does this mean and where can I start troubleshooting? Or is the Visual Graph buggy?

Related

SPARQL query against the ontology in ontop-vkg

Is it possible, to query the ontology alongside a virtual graph? I built a test project (Data tables in PostgreSQL - Ontology and mappings in Protege/w Ontop Plugin), and while querying via the Ontop Sparql Tab works fine, the answers only include triplets from the VKG, but triplets from the ontology do not show up. I tried the same in graphDB with an ontop style repository to no avail.
Thank you in advance!
Attila

Mount a SPARQL endpoint for use with custom ontologies and triple RDFs

I've been trying to figure out how to mount a SPARQL endpoint for a couple of days, but as much as I read I can not understand it.
Comment my intention: I have an open data server mounted on CKAN and my goal is to be able to use SPARQL queries on the data. I know I could not do it directly on the datasets themselves, and I would have to define my own OWL and convert the data I want to use from CSV format (which is the format they are currently in) to RDF triple format (to be used as linked data).
The idea was to first test with the metadata of the repositories that can be generated automatically with the extension ckanext-dcat, but is that I really do not find where to start. I've searched for information on how to install a Virtuoso server for the SPARQL, but the information I've found leaves a lot to be desired, not to say that I can find nowhere to explain how I could actually introduce my own OWLs and RDFs into Virtuoso itself.
Someone who can lend me a hand to know how to start? Thank you
I'm a little confused. Maybe this is two or more questions?
1. How to convert tabular data, like CSV, into the RDF semantic format?
This can be done with an R2RML approach. Karma is a great GUI for that purpose. Like you say, a conversion like that can really be improved with an underlying OWL ontology. But it can be done without creating a custom ontology, too.
I have elaborated on this in the answer to another question.
2. Now that I have some RDF formatted data, how can I expose it with a SPARQL endpoint?
Virtuoso is a reasonable choice. There are multiple ways to deploy it and multiple ways to load the data, and therefore LOTs of tutorial on the subject. Here's one good one, from DBpedia.
If you'd like a simpler path to starting an RDF triplestore with a SPARQL endpoint, Stardog and Blazegraph are available as JARs, and RDF4J can easily be deployed within a container like Tomcat.
All provide web-based graphical interfaces for loading data and running queries, in addition to SPARQL REST endpoints. At least Stardog also provides command-line tools for bulk loading.

Importing ontology from Protege to Sesame

I have created an ontology where I've imported a couple of external ontologies and mapped some of my classes to those in them. When I open the ontology in Protege, I see the classes that I created as well as the mapping to other ontologies.
However, when I open the ontology in a text editor or export it to Sesame, I only see the classes I created or mapped to those in existing ontologies. The ones I haven't touched do not show up. I need all the classes from the imported ontologies for my application. How do I get them to show up in my owl file?
UPDATE1
When I import these ontologies directly into Sesame, all the triples show up. This is in contrast to the case of importing these ontologies to Protege and then uploading the file to Sesame. In the latter case, the classes I haven't modified/mapped to those in my ontology don't show up.
What explains the difference in the behavior?
These are the ontologies I am concerned with:
http://svn.code.sf.net/p/vaccineontology/code/trunk/src/ontology/VO.owl
http://ovae.googlecode.com/svn/trunk/src/ontology/ovae.owl

Linking my own ontology with DBPedia's populated ontology

I have this assignment where I have created an rdfs ontology with artists, their songs, their genres, etc using Protégé and populated it with some data. I am now supposed to connect my ontology and it's data with DBPedia's, so that I can SPARQL for my artists' birthdate. I have successfully imported http://dbpedia.org/ontology/ and http://dbpedia.org/property/ into protégé, but I can't find a way to import the data (http://dbpedia.org/resource/).
Can anyone point me in the right direction?
DBpedia datasets are stored and available for download here (http://wiki.dbpedia.org/Downloads2015-04#h96554-1) but Protégé isn't going to perform well (or at all) when you load a dataset of this size into it.
You may be better downloading various entities by hand such as - http://dbpedia.org/data/Blink-182.ntriples and then loading into Protégé. You can go to any DBPedia page. E.G. http://dbpedia.org/page/Ras_Barker and find download links at the bottom.

semantic web + linked data integration

i'm new to semantic web.
I'm trying to do a sample application where i can query data from different data sources in one query.
i have created a small rdf file which contains references to dbpedia resources for defining localities. my question is : how can i get the data contained in my file and other information which is in the description of the distant resource (for example : the name of the person from the local file, and the total poulation in a city dbpedia-owl:populationTotal from the distant rdf file).
i don't really understand the sparql query language, i tried to use the JENA ARQ API with the SERVICE keyword but it doesn't solve the problem.
Any help please?
I guess you are looking for something like the Semantic Web Client Library, which tries to leverage the GGG. Albeit, the standard exploration algorithm of this framework is that it follows rdfs:seeAlso links. Nevertheless, the general approach seems to be what your are looking for, i.e., you would create a local graph that consists of your seed graph and that traverse the relations up to a certain level, e.g., three steps, resolves the URIs and load that content into your local triple. Utilising advanced technologies like SPARQL federation might be something for later ;)
I have retrived data from two different sources using SPARQL query with named graphs.
I used jena-ARQ to execute the sparql query.