Uploaded image for project: 'Blazegraph (by SYSTAP)'
  1. Blazegraph (by SYSTAP)
  2. BLZG-1092

bigdata stucking on a macbookpro 3GHZ 16GB SSD

    XMLWordPrintable

    Details

    • Type: Bug
    • Status: Closed - Won't Fix
    • Resolution: Incomplete
    • Affects Version/s: BIGDATA_RELEASE_1_3_3
    • Fix Version/s: None
    • Component/s: B+Tree
    • Labels:
      None

      Description

      Hi !

      My targe is to load DBPedia into a namespace of bigdata on a mac book pro 3GHZ/16GB + fast SSD storage.

      The JAGA_OPTS used are

      JAVA_OPTS='-XX +UseG1GC -XX:+UseMembar -server -Xmx10G -XX:MaxDirectMemorySize=6000m'

      The namespace is "triples with Fulltextindex and inferencing:

      com.bigdata.namespace.kb.spo.com.bigdata.btree.BTree.branchingFactor 1024
      com.bigdata.relation.container DBPedia
      com.bigdata.journal.AbstractJournal.bufferMode DiskRW
      com.bigdata.journal.AbstractJournal.file bigdata.jnl
      com.bigdata.journal.AbstractJournal.initialExtent 209715200
      com.bigdata.rdf.store.AbstractTripleStore.vocabularyClass com.bigdata.rdf.vocab.DefaultBigdataVocabulary
      com.bigdata.rdf.store.AbstractTripleStore.textIndex true
      com.bigdata.btree.BTree.branchingFactor 128
      com.bigdata.namespace.kb.lex.com.bigdata.btree.BTree.branchingFactor 400
      com.bigdata.rdf.store.AbstractTripleStore.axiomsClass com.bigdata.rdf.axioms.OwlAxioms
      com.bigdata.service.AbstractTransactionService.minReleaseAge 1
      com.bigdata.rdf.sail.truthMaintenance true
      com.bigdata.journal.AbstractJournal.maximumExtent 209715200
      com.bigdata.rdf.sail.namespace DBPedia
      com.bigdata.relation.class com.bigdata.rdf.store.LocalTripleStore
      com.bigdata.rdf.store.AbstractTripleStore.quads false
      com.bigdata.search.FullTextIndex.fieldsEnabled false
      com.bigdata.relation.namespace DBPedia
      com.bigdata.btree.writeRetentionQueue.capacity 8000
      com.bigdata.rdf.store.AbstractTripleStore.statementIdentifiers false

      AutoCommit set to 300'000 lines, all following the information in
      PerformanceOptimization?(General background on performance tuning for bigdata)
      IOOptimization?(How to tune IO performance, SSD, branching factors, etc.)

      The files are ntriples files, automatically split at 3'000'000 Lines each.

      1. What happened:
      While at the beginning the machine responded quite well with an average load and a speed of ca. 3000000 triples in 143 seconds, after 5 hours and after having processed less then 100'000 Triples the system came to stuck without warnings nor errors. What is happening? and what shell I do (please)?

      Yesterday I uploaded 143 millions of triples in the default namespace kb, then I stopped that because no inferencing was checked. But know it seems it was a bad idea to activate inferencing.

      Thank you for any hint on this matter.

      2. I could now switch back to my other journal with 143 millions triples to continue with, but I would like to have the triples inside an own namespace. I think a sparql update on that would last for months... What could I do in order to reuse/rename a namespace with such noumerous triples?

      Thanks
      Goodbye

      Fabio

        Attachments

          Activity

            People

            Assignee:
            thompsonbry thompsonbry
            Reporter:
            fabioricci fabioricci
            Votes:
            0 Vote for this issue
            Watchers:
            0 Start watching this issue

              Dates

              Created:
              Updated:
              Resolved: