apache / Hadoop Hdfs
Licence: apache-2.0
Mirror of Apache Hadoop HDFS
Stars: ✭ 152
Programming Languages
java
68154 projects - #9 most used programming language
Labels
Projects that are alternatives of or similar to Hadoop Hdfs
Datax
DataX is an open source universal ETL tool that support Cassandra, ClickHouse, DBF, Hive, InfluxDB, Kudu, MySQL, Oracle, Presto(Trino), PostgreSQL, SQL Server
Stars: ✭ 116 (-23.68%)
Mutual labels: hadoop
Spydra
Ephemeral Hadoop clusters using Google Compute Platform
Stars: ✭ 128 (-15.79%)
Mutual labels: hadoop
Drill
Apache Drill is a distributed MPP query layer for self describing data
Stars: ✭ 1,619 (+965.13%)
Mutual labels: hadoop
Hadoopcryptoledger
Hadoop Crypto Ledger - Analyzing CryptoLedgers, such as Bitcoin Blockchain, on Big Data platforms, such as Hadoop/Spark/Flink/Hive
Stars: ✭ 126 (-17.11%)
Mutual labels: hadoop
Gaffer
A large-scale entity and relation database supporting aggregation of properties
Stars: ✭ 1,642 (+980.26%)
Mutual labels: hadoop
Hdfs Shell
HDFS Shell is a HDFS manipulation tool to work with functions integrated in Hadoop DFS
Stars: ✭ 117 (-23.03%)
Mutual labels: hadoop
Parquet4s
Read and write Parquet in Scala. Use Scala classes as schema. No need to start a cluster.
Stars: ✭ 125 (-17.76%)
Mutual labels: hadoop
Ibis
A pandas-like deferred expression system, with first-class SQL support
Stars: ✭ 1,630 (+972.37%)
Mutual labels: hadoop
Airflow Pipeline
An Airflow docker image preconfigured to work well with Spark and Hadoop/EMR
Stars: ✭ 128 (-15.79%)
Mutual labels: hadoop
Spark With Python
Fundamentals of Spark with Python (using PySpark), code examples
Stars: ✭ 150 (-1.32%)
Mutual labels: hadoop
Aliyun Emapreduce Datasources
Extended datasource support for Spark/Hadoop on Aliyun E-MapReduce.
Stars: ✭ 132 (-13.16%)
Mutual labels: hadoop
This project does not contain a readme.
Note that the project description data, including the texts, logos, images, and/or trademarks,
for each open source project belongs to its rightful owner.
If you wish to add or remove any projects, please contact us at [email protected].