Giter VIP home page Giter VIP logo

ldbc_graphalytics's People

Contributors

alexaverbuch avatar amusaafir avatar erackron avatar gillesmag avatar hegyibalint avatar mihaic avatar stijnh avatar szarnyasg avatar thegeman avatar wlngai avatar

Stargazers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

Watchers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

ldbc_graphalytics's Issues

Add EPS in report

Edges per seconds = number of edges / runtime.

We will examine (semantical) TEPS in the future.

Unnesscary duo jars in C++ platforms.

For Java platforms, two jar files is needed, one for default, another for Granula.However, for C++ platforms, it is entirely unnecessary to have two jar files (they are identical).

graph files are not accessible. please provide a mirror

Hello

I should have reported this problem to an administrator, but I did not find any contact on the project page nor on the github space. The server at atlarge.ewi.tudelft.nl which hosts graph files is down for several days. I would appreciate if you resolve the server issue or provide a secondary mirror for accessing those files.

Best regards,
Fadishei

Add an overall validation indicator

As far as I can tell, the only way to determine if there were validation failures in a Graphalytics run is to examine each benchmark validation result. It would be useful to have an indication that the validation succeeded overall, e.g., "Validation successful for all benchmarks" printed at the end of the command line output and at the top of the experimental tab in the HTML report.

Check output directory before running benchmark

It's a waste of time to run the benchmark and see this at the end:

[main] ERROR nl.tudelft.graphalytics.Graphalytics - Failed to write report:
java.io.IOException: Output directory of report is non-empty: "neo4j-report".

Inaccurate timeout on benchmark run.

Graphalytics benchmark launches independent JVM to execute each benchmark run, such that time-out can be guaranteed.

However, the timeout includes now the proprocess, postprocess and validation time, which is not an accurate measurement.

Zeta and Geometric Generators

Hello,

in the paper, you mention the following:

"To support the ability to generate graphs of different characteristics, we have extended Datagen with the capability to dynamically reproduce different distributions by means of plugins. We have already implemented those for the Zeta and Geometric distribution models, but more will be added in the
future as more real graphs are analysed."

Where can I find the modified LDBC graph gens for Zeta and Geometric distributions?

Unclassified states/failure of benchmark run.

For any benchmark, the chance is high that some benchmark runs will fail. It is easier to diagnose the failures if there are well-classified states of benchmark run, e.g.,

initialization
execution
completion
validation

Harmonize naming of SSSP

Exception during report generation with misconfiguration

Exception in Thymeleaf when running benchmark with an algorithm that is not yet configured for all graphs. No result is added for (graph, algorithm) pair, but there should be a "not completed"/"not started" result. Exception text:

Exception evaluating OGNL expression: "report.getResult(graph, algorithm).completedSuccessfully"

Translate all paths to absolute paths

Some paths are not converted correctly by the core to absolute paths before they are passed to the platform extensions. To do: find out which paths, and fix them.

Fix test failures in MR

Restructure project

Graphalytics will only include core and validation (currently named graphalytics-tests). Each platform in platforms will be a separate repository. Platform configuration files will be extracted from config.
Each platform will also be a separate Maven project, depending on the core project. Once packaged, each platform code will be used by the core code for actual benchmarking. run-benchmark.sh will sill be used for running the benchmark.

Obtaining binaries will work as follows:

  1. compile-benchmark.sh without flags will install core and validation in the local Maven repository.
  2. compile-platform.sh (new script) in each platform to obtain platform binaries.
  3. Configure path for platform binaries in graphalytics.
  4. package-benchmark.sh (new script) to create a redistributable archive of core (validation?) and platforms, as well as run-benchmark.sh and platform-specific prepare-benchmark.sh scripts.

Unimplemented simplified validation

A full validation has been deployed on each benchmark run, which is quite heavy and verbose. Sometimes validation dataset might not be available. It will be easier to determine whether a benchmark execution has completed by checking if the output has the correct size.

Continuous Integration

E.g., Jenkins @ TUDelft. At least compilation + unit tests, perhaps integration tests.

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    ๐Ÿ–– Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. ๐Ÿ“Š๐Ÿ“ˆ๐ŸŽ‰

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google โค๏ธ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.