SORT BY Relevancy
Hadoop is a top-level Apache project that provides a Java software framework for storing, managing, processing and analysing the massive datasets produced by enterprise web and cloud computing applications. Inspired by Google’s internal software framework and file system, Hadoop includes the HBase distributed database, HDFS distributed file system, MapReduce application framework, Hive data warehousing application and a number of other open source tools, languages and common utilities.
The Best Open Source for Big Data http://bigdatastudio.com/2013/09/01/the-best-of-open-source-for-big-data/
OpenBTS® is an open-source Unix application that uses a software defined radio (like the USRP) to present a GSM air interface to standard GSM handset and uses a software PBX (Asterisk or Freeswitch) to connect calls. The combination of the ubiquitous GSM air interface with VoIP backhaul could form t
Big data and Hadoop have received a lot of press lately. For all the hype, Hadoop does look like an attractive, low-cost solution that your company could leverage. Aside from limited knowledge on how the technology works, a primary reason organizations struggle to achieve value from big data is lack of a compelling business use case.
The team at MetaScale has given this common roadblock plenty of consideration ? first while making big data decisions within Sears Holdings and now as a "big data accelerator" helping clients find the right answers. Tune in for this A2 Radio broadcast as Ankur Gupta, a big data director at Sears and head of sales, marketing, and operations for MetaScale, shares tips on:
Making a case for Hadoop in your organization
Enabling business analytics by speeding up data workloads
Avoiding common missteps and roadblocks
Developing a clearly defined, long-term big data strategy
Learn of the increased focus placed on data security, compliance, and privacy to deal with the constant threats to enterprises, and how a leading industry actor effectively makes data protection possible in their growing diverse IT ecosystem. As company's environments become more open and complex, sensitive data can be found not only on internal databases, but also in datawarehouses, Big Data (Hadoop or NoSQL) platforms, and file systems including those outsourced and in the Cloud. We will be discussing best practices approaches (like data activity monitoring and SIEM) to make data protection and regulation compliance a reality for the organization, while keeping costs down in an ever growing and changing IT deployment.
The Apache Hadoop software library is a framework that allows for the distributed processing of large data sets across clusters of computers using simple programming models. It is designed to scale up from single servers to thousands of machines, each offering local computation and storage. Rather than rely on hardware to deliver high-avaiability, the library itself is designed to detect and handle failures at the application layer, so delivering a highly-availabile service on top of a cluster of computers, each of which may be prone to failures. Cloudera's open-source Apache Hadoop distribution, CDH , targets enterprise-class deployments of that technology
Learn how to lay down the technology foundation for your analytics transformation. In this session, we'll walk you through the relative advantages and disadvantages for midsized firms of public cloud, private cloud, HPC and clusters, scaling up vs. scaling out, placing data close to compute resources, and more. Find out how midsized firms can get the most from Hadoop, Flash storage, and other tools. We'll introduce you to the infrastructure requirements of big-data, and give you a solid idea of the infrastructure choices midmarket firms have for different types of big-data problems.
Join Host Live Chats
There are no live chats in progress