All Projects → rakam-io → Rakam Api

rakam-io / Rakam Api

Licence: agpl-3.0
📈 Collect customer event data from your apps. (Note that this project only includes the API collector, not the visualization platform)

Programming Languages

java
68154 projects - #9 most used programming language

Projects that are alternatives of or similar to Rakam Api

Spark With Python
Fundamentals of Spark with Python (using PySpark), code examples
Stars: ✭ 150 (-80.57%)
Mutual labels:  analytics, big-data
Sciblog support
Support content for my blog
Stars: ✭ 694 (-10.1%)
Mutual labels:  analytics, big-data
Fili
Easily make RESTful web services for time series reporting with Big Data analytics engines like Druid and SQL Databases.
Stars: ✭ 151 (-80.44%)
Mutual labels:  analytics, big-data
Countly Sdk Cordova
Countly Product Analytics SDK for Cordova, Icenium and Phonegap
Stars: ✭ 69 (-91.06%)
Mutual labels:  analytics, big-data
Crate
CrateDB is a distributed SQL database that makes it simple to store and analyze massive amounts of data in real-time.
Stars: ✭ 3,254 (+321.5%)
Mutual labels:  analytics, big-data
Logisland
Scalable stream processing platform for advanced realtime analytics on top of Kafka and Spark. LogIsland also supports MQTT and Kafka Streams (Flink being in the roadmap). The platform does complex event processing and is suitable for time series analysis. A large set of valuable ready to use processors, data sources and sinks are available.
Stars: ✭ 97 (-87.44%)
Mutual labels:  analytics, big-data
Hyperspace
An open source indexing subsystem that brings index-based query acceleration to Apache Spark™ and big data workloads.
Stars: ✭ 246 (-68.13%)
Mutual labels:  analytics, big-data
Data Science Live Book
An open source book to learn data science, data analysis and machine learning, suitable for all ages!
Stars: ✭ 193 (-75%)
Mutual labels:  analytics, big-data
Trino
Official repository of Trino, the distributed SQL query engine for big data, formerly known as PrestoSQL (https://trino.io)
Stars: ✭ 4,581 (+493.39%)
Mutual labels:  analytics, big-data
awesome-AI-kubernetes
❄️ 🐳 Awesome tools and libs for AI, Deep Learning, Machine Learning, Computer Vision, Data Science, Data Analytics and Cognitive Computing that are baked in the oven to be Native on Kubernetes and Docker with Python, R, Scala, Java, C#, Go, Julia, C++ etc
Stars: ✭ 95 (-87.69%)
Mutual labels:  big-data, analytics
Data Science Career
Career Resources for Data Science, Machine Learning, Big Data and Business Analytics Career Repository
Stars: ✭ 630 (-18.39%)
Mutual labels:  analytics, big-data
Beeva Best Practices
Best Practices and Style Guides in BEEVA
Stars: ✭ 335 (-56.61%)
Mutual labels:  analytics, big-data
Qcportal
A client interface to the QCArchive Project (read-only image of QCFractal)
Stars: ✭ 29 (-96.24%)
Mutual labels:  analytics, big-data
Maha
A framework for rapid reporting API development; with out of the box support for high cardinality dimension lookups with druid.
Stars: ✭ 101 (-86.92%)
Mutual labels:  analytics, big-data
Dremio Oss
Dremio - the missing link in modern data
Stars: ✭ 862 (+11.66%)
Mutual labels:  analytics, big-data
Clickhouse
ClickHouse® is a free analytics DBMS for big data
Stars: ✭ 21,089 (+2631.74%)
Mutual labels:  analytics, big-data
Delta
An open-source storage layer that brings scalable, ACID transactions to Apache Spark™ and big data workloads.
Stars: ✭ 3,903 (+405.57%)
Mutual labels:  analytics, big-data
Pachyderm
Reproducible Data Science at Scale!
Stars: ✭ 5,305 (+587.18%)
Mutual labels:  analytics, big-data
Cython
The most widely used Python to C compiler
Stars: ✭ 6,588 (+753.37%)
Mutual labels:  big-data
Sdc
Intel® Scalable Dataframe Compiler for Pandas*
Stars: ✭ 623 (-19.3%)
Mutual labels:  big-data

Build Status Deploy

Rakam

Rakam is an analytics platform that allows you to create your analytics services.

Features / Goals

Rakam is a modular analytics platform that gives you a set of features to create your own analytics service.

Typical workflow of using Rakam:

We also provide user interface for Rakam as a separate product called Rakam UI. You can create custom reports with SQL, dashboards, funnel and retention reports via Rakam UI.

All these features come with a single box, you just need to specify which modules you want to use using a configuration file (config.properties) and Rakam will do the rest for you. We also provide cloud deployment tools for scaling your Rakam cluster easily.

Deployment

If your event data-set can fit in a single server, we recommend using Postgresql backend. Rakam will collect all your events in row-oriented format in a Postgresql node. All the features provided by Rakam are supported in Postgresql deployment type. Please note that we support Postgresql 11 because we're using new features such as partitioning and BRIN indexes for performance.

However Rakam is designed to be highly scalable in order to provide a solution for high work-loads. You can configure Rakam to send events to a distributed commit-log such as Apache Kafka or Amazon Kinesis in serialized Apache Avro format and process data in PrestoDB workers and store them in a distributed filesystem in a columnar format.

Heroku

You can deploy Rakam to Heroku using Heroku button, it uses Heroku Postgresql add-on for your app and uses Postgresql deployment type.

Deploy

Docker

Run the following command to start a Postgresql server in docker container and Rakam API in your local environment.

docker run -d --name rakam-db -e POSTGRES_PASSWORD=dummy -e POSTGRES_USER=rakam postgres:10.1 && docker run --link rakam-db --name rakam -p 9999:9999 -e RAKAM_CONFIG_LOCK__KEY=mylockKey -e RAKAM_CONFIG_STORE_ADAPTER_POSTGRESQL_URL=postgres://rakam:[email protected]:5432/rakam buremba/rakam

After docker container is started, visit http://127.0.0.1:9999 and follow the instructions. You can also register your local Rakam API to Rakam BI at http://app.rakam.io or directly use Rakam API. You may also consult to API documentation for details of the API.

We also provide a docker-compose definition for a Postgresql backend. Create a docker-compose.yml with this definition and run the command docker-compose -f docker-compose.yml up -d.

version: '2.1'
services:
  rakam-db:
    image: postgres:11.4
    environment:
      - POSTGRES_PASSWORD=dummy
      - POSTGRES_USER=rakam
    healthcheck:
      test: ["CMD-SHELL", "pg_isready"]
      interval: 5s
      timeout: 5s
      retries: 3
  rakam-api:
    image: buremba/rakam
    environment:
      - RAKAM_CONFIG_STORE_ADAPTER_POSTGRESQL_URL=postgres://rakam:[email protected]:5432/rakam
      - RAKAM_CONFIG_LOCK__KEY=mylockKey
    ports:
      - "9999:9999"
    depends_on:
      rakam-db:
        condition: service_healthy

You can set config variables for Rakam instance using environment variables. All properties in config.properties file can be set via environment variable RAKAM_CONFIG_property_name_dots_replaced_by_underscore. For example, if you want to set store.adapter=postgresql you need to set environment variable RAKAM_CONFIG_STORE_ADAPTER=postgresql. Also the dash - is replaced by double underscore character __. Therefore the environment variable RAKAM_CONFIG_LOCK__KEY corresponds to lock-key config property.

Dockerfile will generate config.properties file from environment variables in docker container that start with RAKAM_CONFIG prefix.

In order to set environment variables for container, you may use -e flag for for docker run but we advice you to set all environment variables in a file and use --env-file flag when starting your container.

Then you can share same file among the Rakam containers. If Dockerfile can't find any environment variable starts with RAKAM_CONFIG, it tries to connect Postgresql instance created with docker-compose.

AWS (Terraform)

See https://github.com/rakam-io/rakam-api-terraform-aws.

Terraform installer is the recommended way to deploy Rakam in production because it automatically handles most of the complexity like fail over and load-balancing.

Custom

  • Download Java 1.8 for your operating system.
  • Download latest version from Bintray ([VERSION]/rakam-[VERSION]-.bundle.tar.gz) extract package.
  • Modify etc/config.properties (sample for Postgresql deployment type) file and run bin/launcher start.
  • The launcher script can take the following arguments: start|restart|stop|status|run. bin/launcher run will start Rakam in foreground.

Building Rakam

You can try the master branch by pulling the source code from Github and building Rakam using Maven:

Requirements
  • Java 8
  • Maven 3.2.3+ (for building)
git clone https://github.com/rakam-io/rakam.git
cd rakam
mvn clean install package -DskipTests
Running the application locally
rakam/target/rakam-*-bundle/rakam-*/bin/launcher.py run --config rakam/target/rakam-*-bundle/rakam-*/etc/config.properties

Note that you need to modify config.properties file in order to be able to start Rakam. (sample for Postgresql deployment type)

Running Rakam in your IDE

Since we already use Maven, you can import Rakam to your IDE using the root pom.xml file. We recommend using Intellij IDEA since the core team uses it when developing Rakam. Here is a sample configuration for executing Rakam in your IDE:

Main Class: org.rakam.ServiceStarter
VM Options: -ea -Xmx2G -Dconfig=YOUR_CONFIG_DIRECTORY/config.properties
Working directory: $MODULE_DIR$
Use classpath of module: rakam

Managed

We're also working for managed Rakam cluster, we will deploy Rakam to our AWS accounts and manage it for you so that you don't need to worry about scaling, managing and software updates. We will do it for you. Please shoot us an email to [email protected] if you want to test our managed Rakam service.

Web application

This repository contains Rakam API server that allows you to interact with Rakam using a REST interface. If you already have a frontend and developed a custom analytics service based on Rakam, it's all you need.

However, we also developed Rakam Web Application that allows you to analyze your user and event data-set but performing SQL queries, visualising your data in various charts, creating (real-time) dashboards and custom reports. You can turn Rakam into a analytics web service similar to Mixpanel, Kissmetrics and Localytics using the web application. Otherwise, Rakam server is similar to Keen.io with SQL as query language and some extra features.

Another nice property of Rakam web application is being BI (Business Intelligence) tool. If you can disable collect APIs and connect Rakam to your SQL database with JDBC adapter and use Rakam application to query your data in your database. Rakam Web Application has various charting formats, supports parameterized SQL queries, custom pages that allows you to design pages with internal components.

Contribution

Currently I'm actively working on Rakam. If you want to contribute the project or suggest an idea feel free to fork it or create a ticket for your suggestions. I promise to respond you ASAP. The purpose of Rakam is being generic data analysis tool which can be a solution for many use cases. Rakam still needs too much work and will be evolved based on people's needs so your thoughts are important.

Acknowledgment

YourKit

We use YourKit Java Profiler in order to monitor the JVM instances for identifing the bugs and potential bottlenecks. Kudos to YourKit for supporting Rakam with your full-featured Java Profile!

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].