fastdata-clusterFast Data Cluster (Apache Cassandra, Kafka, Spark, Flink, YARN and HDFS with Vagrant and VirtualBox)
Stars: ✭ 20 (-84.37%)
leaflet heatmap简单的可视化湖州通话数据 假设数据量很大,没法用浏览器直接绘制热力图,把绘制热力图这一步骤放到线下计算分析。使用Apache Spark并行计算数据之后,再使用Apache Spark绘制热力图,然后用leafletjs加载OpenStreetMap图层和热力图图层,以达到良好的交互效果。现在使用Apache Spark实现绘制,可能是Apache Spark不擅长这方面的计算或者是我没有设计好算法,并行计算的速度比不上单机计算。Apache Spark绘制热力图和计算代码在这 https://github.com/yuanzhaokang/ParallelizeHeatmap.git .
Stars: ✭ 13 (-89.84%)
Spark With PythonFundamentals of Spark with Python (using PySpark), code examples
Stars: ✭ 150 (+17.19%)
ElasticlusterCreate clusters of VMs on the cloud and configure them with Ansible.
Stars: ✭ 298 (+132.81%)
AkkeeperAn easy way to deploy your Akka services to a distributed environment.
Stars: ✭ 30 (-76.56%)
fsbrowserFast desktop client for Hadoop Distributed File System
Stars: ✭ 27 (-78.91%)
ros hadoopHadoop splittable InputFormat for ROS. Process rosbag with Hadoop Spark and other HDFS compatible systems.
Stars: ✭ 92 (-28.12%)
Jsr203 HadoopA Java NIO file system provider for HDFS
Stars: ✭ 35 (-72.66%)
terasliceScalable data processing pipelines in JavaScript
Stars: ✭ 48 (-62.5%)
Bigdata💎🔥大数据学习笔记
Stars: ✭ 488 (+281.25%)
IbisA pandas-like deferred expression system, with first-class SQL support
Stars: ✭ 1,630 (+1173.44%)
Devops Python Tools80+ DevOps & Data CLI Tools - AWS, GCP, GCF Python Cloud Function, Log Anonymizer, Spark, Hadoop, HBase, Hive, Impala, Linux, Docker, Spark Data Converters & Validators (Avro/Parquet/JSON/CSV/INI/XML/YAML), Travis CI, AWS CloudFormation, Elasticsearch, Solr etc.
Stars: ✭ 406 (+217.19%)
waspWASP is a framework to build complex real time big data applications. It relies on a kind of Kappa/Lambda architecture mainly leveraging Kafka and Spark. If you need to ingest huge amount of heterogeneous data and analyze them through complex pipelines, this is the framework for you.
Stars: ✭ 19 (-85.16%)
datasqueezeHadoop utility to compact small files
Stars: ✭ 18 (-85.94%)
aaocp一个对用户行为日志进行分析的大数据项目
Stars: ✭ 53 (-58.59%)
Hadoop For GeoeventArcGIS GeoEvent Server sample Hadoop connector for storing GeoEvents in HDFS.
Stars: ✭ 5 (-96.09%)
Bigdata Interview🎯 🌟[大数据面试题]分享自己在网络上收集的大数据相关的面试题以及自己的答案总结.目前包含Hadoop/Hive/Spark/Flink/Hbase/Kafka/Zookeeper框架的面试题知识总结
Stars: ✭ 857 (+569.53%)
Repository个人学习知识库涉及到数据仓库建模、实时计算、大数据、Java、算法等。
Stars: ✭ 92 (-28.12%)
docker-hadoopDocker image for main Apache Hadoop components (Yarn/Hdfs)
Stars: ✭ 59 (-53.91%)
bigdata-funA complete (distributed) BigData stack, running in containers
Stars: ✭ 14 (-89.06%)
py-hdfs-mountMount HDFS with fuse, works with kerberos!
Stars: ✭ 13 (-89.84%)
God Of Bigdata专注大数据学习面试,大数据成神之路开启。Flink/Spark/Hadoop/Hbase/Hive...
Stars: ✭ 6,008 (+4593.75%)
hive to es同步Hive数据仓库数据到Elasticsearch的小工具
Stars: ✭ 21 (-83.59%)
CamusMirror of Linkedin's Camus
Stars: ✭ 81 (-36.72%)
Wifi基于wifi抓取信息的大数据查询分析系统
Stars: ✭ 93 (-27.34%)
DynamometerA tool for scale and performance testing of HDFS with a specific focus on the NameNode.
Stars: ✭ 122 (-4.69%)
SwarmletA self-hosted, open-source Platform as a Service that enables easy swarm deployments, load balancing, automatic SSL, metrics, analytics and more.
Stars: ✭ 373 (+191.41%)
init ec2init EC2 cluster, for free-password-login(ubuntu and root). for hostname, for hosts file.
Stars: ✭ 11 (-91.41%)
CloudbreakA tool for provisioning and managing Apache Hadoop clusters in the cloud. Cloudbreak, as part of the Hortonworks Data Platform, makes it easy to provision, configure and elastically grow HDP clusters on cloud infrastructure. Cloudbreak can be used to provision Hadoop across cloud infrastructure providers including AWS, Azure, GCP and OpenStack.
Stars: ✭ 301 (+135.16%)
Hdfs ShellHDFS Shell is a HDFS manipulation tool to work with functions integrated in Hadoop DFS
Stars: ✭ 117 (-8.59%)
Tempsλ A selfhostable serverless function runtime. Inspired by zeit now.
Stars: ✭ 15 (-88.28%)
pmml4sPMML scoring library for Scala
Stars: ✭ 49 (-61.72%)
corcAn ORC File Scheme for the Cascading data processing platform.
Stars: ✭ 14 (-89.06%)
big-data-exploration[Archive] Intern project - Big Data Exploration using MongoDB - This Repository is NOT a supported MongoDB product
Stars: ✭ 43 (-66.41%)
push-package-action| Public | GitHub Action to Push a Package to Octopus Deploy
Stars: ✭ 23 (-82.03%)
easy qsubEasily submitting multiple PBS jobs or running local jobs in parallel. Multiple input files supported.
Stars: ✭ 26 (-79.69%)
LogAnalyzeHelper论坛日志分析系统清洗程序(包含IP规则库,UDF开发,MapReduce程序,日志数据)
Stars: ✭ 33 (-74.22%)
create-release-action| Public | GitHub Action to Create a Release in Octopus Deploy
Stars: ✭ 68 (-46.87%)
deploy shard mongodbThis repository has a set of scripts and resources required for deploying MongoDB replicated sharded cluster.
Stars: ✭ 17 (-86.72%)
LibrarianEasily host your iOS and Android builds locally!
Stars: ✭ 35 (-72.66%)
managerThe API endpoint that manages nebula orchestrator clusters
Stars: ✭ 28 (-78.12%)
testnet deployDeployment scripts and monitoring configuration for a Cosmos Validator setup
Stars: ✭ 19 (-85.16%)
ML-CaPsuleML-capsule is a Project for beginners and experienced data science Enthusiasts who don't have a mentor or guidance and wish to learn Machine learning. Using our repo they can learn ML, DL, and many related technologies with different real-world projects and become Interview ready.
Stars: ✭ 177 (+38.28%)
laniakeaLaniakea is a utility for managing instances at various cloud providers and aids in setting up a fuzzing cluster.
Stars: ✭ 28 (-78.12%)
datalake-etl-pipelineSimplified ETL process in Hadoop using Apache Spark. Has complete ETL pipeline for datalake. SparkSession extensions, DataFrame validation, Column extensions, SQL functions, and DataFrame transformations
Stars: ✭ 39 (-69.53%)
Meteor-Mailer📮 Bulletproof email queue on top of NodeMailer with support of multiple clusters and servers setup
Stars: ✭ 21 (-83.59%)