What are the features and application of HBase?

HBase

HBase is a column-oriented distributed database developed on top of the Hadoop file system. It is an open-source project that may be scaled horizontally. HBase is a data architecture comparable to Google's big table that is meant to allow fast random access to massive volumes of structured data. It makes use of the Hadoop File System's fault tolerance (HDFS).

Apache HBase is a Hadoop -based distributed, scalable NoSqlb Big data storage.HBase is capable of hosting very large tables-billions of rows and millions of columns-and of providing real-time, random read/write access to Hadoop data. HBase is a multi-column data store inspired by Google Bigtable, a database interface to Google's proprietary File System. HBase adds Bigtable-like features to read/write access to Hadoop-compatible file systems like MapR XD. HBase scales linearly over very large datasets and allows for the easy combination of data sources with heterogeneous topologies and schemas. 

 Features of HBase

  • HBase is linearly scalable.
  • It has automatic failure support.
  • It provides consistent reading and writing. 
  • It integrates with Hadoop, both as a source and a destination.
  • It has an easy java API for clients.
  • It provides data replication across clusters.


Where to use. HBase? 

  • Apache HBase is used to have random, real-time read/write access to Big Data. 
  • It hosts very large tables on top of clusters of commodity hardware.
  • Apache HBase is a non-relational database modeled after Google's Bigtable. Bigtable acts up on Google File System, likewise, Apache HBase works on top of Hadoop and HDFS.


Applications of HBase

  • It is used whenever there is a need to develop complex applications.
  • HBase is used whenever we need to provide fast random access to available data. 
  • Companies such as Facebook, Twitter, Yahoo, and Adobe use HBase internally.


Comments

Popular posts from this blog

Suppose that a data warehouse for Big-University consists of the following four dimensions: student, course, semester, and instructor, and two measures count and avg_grade. When at the lowest conceptual level (e.g., for a given student, course, semester, and instructor combination), the avg_grade measure stores the actual course grade of the student. At higher conceptual levels, avg_grade stores the average grade for the given combination. a) Draw a snowflake schema diagram for the data warehouse. b) Starting with the base cuboid [student, course, semester, instructor], what specific OLAP operations (e.g., roll-up from semester to year) should one perform in order to list the average grade of CS courses for each BigUniversity student. c) If each dimension has five levels (including all), such as “student < major < status < university < all”, how many cuboids will this cube contain (including the base and apex cuboids)?

Suppose that a data warehouse consists of the four dimensions; date, spectator, location, and game, and the two measures, count and charge, where charge is the fee that a spectator pays when watching a game on a given date. Spectators may be students, adults, or seniors, with each category having its own charge rate. a) Draw a star schema diagram for the data b) Starting with the base cuboid [date; spectator; location; game], what specific OLAP operations should perform in order to list the total charge paid by student spectators at GM Place in 2004?

Discuss classification or taxonomy of virtualization at different levels.