All Projects → linkedpipes → Etl

linkedpipes / Etl

Licence: other
LinkedPipes ETL is an RDF based, lightweight ETL tool

Programming Languages

java
68154 projects - #9 most used programming language

Projects that are alternatives of or similar to Etl

Grafter
Linked Data & RDF Manufacturing Tools in Clojure
Stars: ✭ 174 (+97.73%)
Mutual labels:  linked-data, rdf, etl
rdf-ldp
A suite of LDP software and middleware for RDF.rb & Rack
Stars: ✭ 14 (-84.09%)
Mutual labels:  linked-data, rdf
Jsonld.js
A JSON-LD Processor and API implementation in JavaScript
Stars: ✭ 1,212 (+1277.27%)
Mutual labels:  linked-data, rdf
jarql
SPARQL for JSON: Turn JSON into RDF using SPARQL syntax
Stars: ✭ 19 (-78.41%)
Mutual labels:  linked-data, rdf
sirene-ld
Transformation du répertoire SIRENE (CSV) au format RDF pour publication en Linked Data
Stars: ✭ 31 (-64.77%)
Mutual labels:  linked-data, rdf
sparql-micro-service
SPARQL micro-services: A lightweight approach to query Web APIs with SPARQL
Stars: ✭ 22 (-75%)
Mutual labels:  linked-data, rdf
LinkedDataHub
The Knowledge Graph notebook. Apache license.
Stars: ✭ 150 (+70.45%)
Mutual labels:  linked-data, rdf
YALC
🕸 YALC: Yet Another LOD Cloud (registry of Linked Open Datasets).
Stars: ✭ 14 (-84.09%)
Mutual labels:  linked-data, rdf
Rdf
RDF.rb is a pure-Ruby library for working with Resource Description Framework (RDF) data.
Stars: ✭ 353 (+301.14%)
Mutual labels:  linked-data, rdf
Semanticmediawiki
🔗 Semantic MediaWiki turns MediaWiki into a knowledge management platform with query and export capabilities
Stars: ✭ 359 (+307.95%)
Mutual labels:  linked-data, rdf
Pyld
JSON-LD processor written in Python
Stars: ✭ 413 (+369.32%)
Mutual labels:  linked-data, rdf
jsonld-streaming-serializer.js
A fast and lightweight streaming JSON-LD serializer for JavaScript
Stars: ✭ 20 (-77.27%)
Mutual labels:  linked-data, rdf
cubiql
CubiQL: A GraphQL service for querying multidimensional Linked Data Cubes
Stars: ✭ 40 (-54.55%)
Mutual labels:  linked-data, rdf
awesome-ontology
A curated list of ontology things
Stars: ✭ 73 (-17.05%)
Mutual labels:  linked-data, rdf
CSV2RDF
Streaming, transforming, SPARQL-based CSV to RDF converter. Apache license.
Stars: ✭ 48 (-45.45%)
Mutual labels:  linked-data, rdf
link-redux
Linked Data Rendering for humans with React
Stars: ✭ 28 (-68.18%)
Mutual labels:  linked-data, rdf
Rdf Dereference.js
Dereference any URL for its RDF contents
Stars: ✭ 18 (-79.55%)
Mutual labels:  linked-data, rdf
rdf2x
RDF2X converts big RDF datasets to the relational database model, CSV, JSON and ElasticSearch.
Stars: ✭ 43 (-51.14%)
Mutual labels:  linked-data, rdf
rdfa-streaming-parser.js
A fast and lightweight streaming RDFa parser for JavaScript
Stars: ✭ 15 (-82.95%)
Mutual labels:  linked-data, rdf
Tropy
Research photo management
Stars: ✭ 337 (+282.95%)
Mutual labels:  linked-data, rdf

LinkedPipes ETL

Build Status

LinkedPipes ETL is an RDF based, lightweight ETL tool.

Requirements

For building locally

Installation and startup

You can run LP-ETL in Docker, or build it from the source.

Docker

To start LP-ETL master branch on http://localhost:8080, you can use a one-liner:

curl https://raw.githubusercontent.com/linkedpipes/etl/master/docker-compose.yml | docker-compose -f - up

Alternatively, you can clone the entire repository

git clone https://github.com/linkedpipes/etl.git

and run

docker-compose up

Note that this uses just the docker-compose.yml file, so the rest of the cloned repository is useless.

You may need to run the commands as sudo or be in the docker group.

Configuration

Each component (executor, executor-monitor, storage, frontend) has separate Dockerfile.

Environment variables:

  • LP_ETL_BUILD_BRANCH - The Dockerfiles are designed to run build from the github repository, the branch is set using this property, default is master.
  • LP_ETL_BUILD_JAVA_TEST - Set to empty to allow to run Java tests, this will slow down the build.
  • LP_ETL_DOMAIN - The URL of the instance, this is used instead of the domain.uri from the configuration.
  • LP_ETL_FTP - The URL of the FTP server, this is used instead of the executor-monitor.ftp.uri from the configuration.

For Docker Compose, there are additional environment variables:

  • LP_ETL_PORT - Specify port mapping for frontend, this is where you can connect to your instance. This does NOT have to be the same as port in LP_ETL_DOMAIN in case of reverse-proxying.

For example to run LP-ETL from develop branch on http://localhost:9080 use can use following command:

curl https://raw.githubusercontent.com/linkedpipes/etl/develop/docker-compose.yml | LP_ETL_PORT=9080 LP_ETL_DOMAIN=http://localhost:9080 docker-compose -f - up

docker-compose utilizes several volumes that can be used to access/provide data. See docker-compose.yml comments for examples and configuration. You may want to create your own docker-compose.yml for custom configuration.

From source on Linux

Installation

$ git clone https://github.com/linkedpipes/etl.git
$ cd etl
$ mvn install

Configuration

The configuration file deploy/configuration.properties can be edited, mainly changing paths to working, storage, log and library directories.

Startup

$ cd deploy
$ ./executor.sh >> executor.log &
$ ./executor-monitor.sh >> executor-monitor.log &
$ ./storage.sh >> storage.log &
$ ./frontend.sh >> frontend.log &

Running LP-ETL as a systemd service

See example service files in the deploy/systemd folder.

From source on Windows

Note that it is also possible to use Bash on Ubuntu on Windows or Cygwin and proceed as with Linux.

Installation

git clone https://github.com/linkedpipes/etl.git
cd etl
mvn install

Configuration

The configuration file deploy/configuration.properties can be edited, mainly changing paths to working, storage, log and library directories.

Startup

In the deploy folder, run

  • executor.bat
  • executor-monitor.bat
  • storage.bat
  • frontend.bat

Plugins - Components

The components live in the jars directory. Detailed description of how to create your own is coming soon, in the meantime, you can copy an existing component and change it.

Update notes

Update note 5: 2019-09-03 breaking changes in the configuration file. Remove /api/v1 from the executor-monitor.webserver.uri, so it loolks like: executor-monitor.webserver.uri = http://localhost:8081. You can also remove executor.execution.uriPrefix as the value is derived from domain.uri.

Update note 4: 2019-07-03 we changed the way frontend is run. If you do not use our script to run it, you need to update yours.

Update note 3: When upgrading from develop prior to 2017-02-14, you need to delete {deploy}/jars and {deploy}/osgi.

Update note 2: When upgrading from master prior to 2016-11-04, you need to move your pipelines folder from e.g., /data/lp/etl/pipelines to /data/lp/etl/storage/pipelines, update the configuration.properites file and possibly the update/restart scripts as there is a new component, storage.

Update note: When upgrading from master prior to 2016-04-07, you need to delete your old execution data (e.g., in /data/lp/etl/working/data)

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].