Hue - The open source Apache Hadoop UI

  •        0

Hue is a Web application for interacting with Apache Hadoop. It supports a FileBrowser for accessing HDFS, JobBrowser for accessing MapReduce jobs (MR1/MR2-YARN), Job Designer for creating MapReduce/Streaming/Java jobs, HBase Browser for exploring and modifying HBase tables and data, Oozie App for submitting and scheduling workflows and bundles, A Pig/HBase/Sqoop2 shell, Beeswax application for executing Hive queries, Search app for querying Solr and Solr Cloud.

http://gethue.com
https://github.com/cloudera/hue

Tags
Implementation
License
Platform

   




Related Projects

hadoop-cdh5 - Vagrant+Chef cookbook to install hadoop, hue, hive and much more (via cloudera repos)


Vagrant+Chef cookbook to install hadoop, hue, hive and much more (via cloudera repos)

Hadoop Common


Apache Hadoop is a framework for running applications on large clusters built of commodity hardware. Hadoop common supports other Hadoop subprojects

Cascalog - Data processing on Hadoop


Cascalog is a fully-featured data processing and querying library for Clojure or Java. The main use cases for Cascalog are processing "Big Data" on top of Hadoop or doing analysis on your local computer. Cascalog is a replacement for tools like Pig, Hive, and Cascading and operates at a significantly higher level of abstraction than those tools.

HPCC System - Hadoop alternative


HPCC is a proven and battle-tested platform for manipulating, transforming, querying and data warehousing Big Data. It supports two type of configuration. Thor is responsible for consuming vast amounts of data, transforming, linking and indexing that data. It functions as a distributed file system with parallel processing power spread across the nodes. Roxie, the Data Delivery Engine, provides separate high-performance online query processing and data warehouse capabilities.

Kylin - Extreme OLAP Engine for Big Data


Apache Kylin is an open source Distributed Analytics Engine designed to provide SQL interface and multi-dimensional analysis (OLAP) on Hadoop supporting extremely large datasets, original contributed from eBay Inc. It is designed to reduce query latency on Hadoop for 10+ billions of rows of data. It offers ANSI SQL on Hadoop and supports most ANSI SQL query functions.

gptools-for-aws - GP Tools for Amazon Web Services Elastic Map Reduce (Hosted Hadoop Framework)


GP Tools for Amazon Web Services Elastic Map Reduce (Hosted Hadoop Framework)

lmapred - Local map-reduce tools like Hadoop streaming


Local map-reduce tools like Hadoop streaming

teratorn - Big data processing on Hadoop for Map of Life.


Big data processing on Hadoop for Map of Life.

jydoop - Efficient Hadoop Map-Reduce in Python


Querying Hadoop/HBase using custom java classes is complicated and tedious. It's very difficult to test and debug analyses on small sets of sample data, or without setting up a Hadoop/HBase cluster.Writing analyses in Python allows for easier local development + testing without having to set up Hadoop or HBase. The same analysis scripts can then be deployed to a production cluster configuration.

Apache Tajo - A big data warehouse system on Hadoop


Apache Tajo is a robust big data relational and distributed data warehouse system for Apache Hadoop. Tajo is designed for low-latency and scalable ad-hoc queries, online aggregation, and ETL (extract-transform-load process) on large-data sets stored on HDFS (Hadoop Distributed File System) and other data sources.

hadoop-mr-examples - Polyglot Map Reduce Examples on Hadoop


Polyglot Map Reduce Examples on Hadoop

i2b2-hadoop - A Hadoop Map/Reduce implementation of processing and executing I2B2 CRC Queries


A Hadoop Map/Reduce implementation of processing and executing I2B2 CRC Queries

hello-hadoop - A simple Hadoop Map/Reduce Job (WordCount)


A simple Hadoop Map/Reduce Job (WordCount)

knowing-hadoop - Hadoop Map-Reduce Job for Knowing


Hadoop Map-Reduce Job for Knowing

spatial-framework-for-hadoop


The __Spatial Framework for Hadoop__ allows developers and data scientists to use the Hadoop data processing system for spatial data analysis.For tools, [samples](https://github.com/Esri/gis-tools-for-hadoop/tree/master/samples), and [tutorials](https://github.com/Esri/gis-tools-for-hadoop/wiki) that use this framework, head over to [GIS Tools for Hadoop](https://github.com/Esri/gis-tools-for-hadoop).

gis-tools-for-hadoop


* [Tutorial: An Introduction for Beginners] (https://github.com/Esri/gis-tools-for-hadoop/wiki/GIS-Tools-for-Hadoop-for-Beginners)* [Tutorial: Aggregating Data Into Bins](https://github.com/Esri/gis-tools-for-hadoop/wiki/Aggregating-CSV-Data-%28Spatial-Binning%29)* [Tutorial: Correcting your ArcGIS Projection](https://github.com/Esri/gis-tools-for-hadoop/wiki/Correcting-Projection-in-ArcGIS)* [Updated Wiki page for the Spatial-Framework-for-Hadoop](https://github.com/Esri/spatial-framework-for-h

Spark - Fast Cluster Computing


Apache Spark is an open source cluster computing system that aims to make data analytics fast — both fast to run and fast to write. To run programs faster, Spark offers a general execution model that can optimize arbitrary operator graphs, and supports in-memory computing, which lets it query data faster than disk-based engines like Hadoop.

map-reduce-console - Browser based map-reduce console to quickly prototype hadoop jobs


Browser based map-reduce console to quickly prototype hadoop jobs

mrjob - Run MapReduce jobs on Hadoop or Amazon Web Services


mrjob is a Python 2.7/3.3+ package that helps you write and run Hadoop Streaming jobs. It fully supports Amazon's Elastic MapReduce (EMR) service, which allows you to buy time on a Hadoop cluster on an hourly basis. mrjob has basic support for Google Cloud Dataproc (Dataproc) which allows you to buy time on a Hadoop cluster on a minute-by-minute basis. It also works with your own Hadoop cluster.