Welcome!

OpenStack Journal Authors: Pat Romanski, Elizabeth White, Matthew McKenna, Liz McMillan, Ed Featherston

Related Topics: @BigDataExpo, Linux Containers, Open Source Cloud, Containers Expo Blog, Server Monitoring, @CloudExpo, Apache, OpenStack Journal

@BigDataExpo: Article

Red Hat Unveils Big Data and Open Hybrid Cloud Direction

Is building a robust network of ecosystem and enterprise integration partners to deliver Big Data solutions

Red Hat on Wednesday announced its Big Data direction and solutions to satisfy enterprise requirements for highly reliable, scalable, and manageable solutions to effectively run their Big Data analytics workloads. In addition, Red Hat announced that the company will contribute its Red Hat Storage Hadoop plug-in to the Apache Hadoop open community to transform Red Hat Storage into a fully supported, Hadoop-compatible file system for Big Data environments, and that Red Hat is building a robust network of ecosystem and enterprise integration partners to deliver comprehensive Big Data solutions to enterprise customers.

Red Hat Big Data infrastructure and application platforms are suited for enterprises leveraging the open hybrid cloud environment. Red Hat is working with the open cloud community to support Big Data customers. Many enterprises worldwide use public cloud infrastructure, such as Amazon Web Services (AWS), for the development, proof-of-concept, and pre-production phases of their Big Data projects. The workloads are then moved to their private clouds to scale up the analytics with the larger data set. An open hybrid cloud environment enables enterprises to transfer workloads from the public cloud into their private cloud without the need to re-tool their applications. Red Hat is actively engaged in the open cloud community through projects like OpenStack and OpenShift Origin to help meet these enterprise Big Data expectations both today and in the future.

There are several Red Hat solutions available to effectively manage enterprise Big Data workloads. Focused on three primary areas, Red Hat's big data direction includes extending its product portfolio to deliver enhanced enterprise-class infrastructure solutions and application platforms, and partnering with leading big data analytics vendors and integrators.

Red Hat's Big Data Infrastructure Solutions

  • Red Hat Enterprise Linux - According to the Jan. 2012 The Linux Foundation Enterprise Linux User Report, the majority of Big Data implementations run on Linux and as the leading provider of commercial Linux1, Red Hat Enterprise Linux is a leading platform for Big Data deployments. Red Hat Enterprise Linux excels in distributed architectures and includes features that address critical big data needs. Managing tremendous data volumes and intensive analytic processing requires an infrastructure designed for high performance, reliability, fine-grained resource management, and scale-out storage. Red Hat Enterprise Linux addresses these challenges while adding the ability to develop, integrate, and secure big data applications reliably and scale easily to keep up with the pace that data is generated, analyzed, or transferred. This can be accomplished in the cloud, making it easier to store, aggregate, normalize, and integrate data from sources across multiple platforms, whether they are deployed as physical, virtual, or cloud-based resources.
  • Red Hat Storage - Built on the trusted Red Hat Enterprise Linux operating system and the proven GlusterFS distributed file system, Red Hat Storage Servers can be used to pool inexpensive commodity servers to provide a cost-effective, scalable, and reliable storage solution for Big Data.

Red Hat intends to make its Hadoop plug-in for Red Hat Storage available to the Hadoop community later this year. Currently in technology preview, the Red Hat Storage Apache Hadoop plug-in provides a new storage option for enterprise Hadoop deployments that delivers enterprise storage features while maintaining the API compatibility and local data access the Hadoop community expects. Red Hat Storage brings enterprise-class features to Big Data environments, such as Geo replication, High Availability, POSIX compliance, disaster recovery, and management, without compromising API compatibility and data locality. Customers now have a unified data and scale out storage software platform to accommodate files and objects deployed across physical, virtual, public and hybrid cloud resources.

  • Red Hat Enterprise Virtualization - Announced in Dec. 2012, Red Hat Enterprise Virtualization 3.1 is integrated with Red Hat Storage, enabling it to access the secure, shared storage pool managed by Red Hat Storage. This integration also offers enterprises reduced operational costs, expanded portability, choice of infrastructure, scalability, availability and the power of community-driven innovation with the contributions of the open source oVirt and Gluster projects. The combination of these platforms furthers Red Hat's open hybrid cloud vision of an integrated and converged Red Hat Storage and Red Hat Enterprise Virtualization node that serves both compute and storage resources.

Red Hat's Big Data Application and Integration Platforms

  • Red Hat JBoss Middleware - Red Hat JBoss Middleware provides enterprises with powerful technologies for creating and integrating big data-driven applications that are able to interact with new and emerging technologies like Hadoop or MongoDB. Big data is only valuable when businesses can extract information and respond intelligently. Red Hat JBoss Middleware solutions can populate large volumes and varieties of data quickly and reliably into Hadoop with high speed messaging technologies; simplify working with MongoDB through Hibernate OGM; process large volumes of data quickly and easily with Red Hat JBoss Data Grid; access Hadoop along with your traditional data sources with JBoss Enterprise Data Services Platform; and identify opportunities and threats through pattern recognition with JBoss Enterprise BRMS. Red Hat's middleware portfolio is well-suited to help enterprises seize the opportunities of big data.

Big Data Partnerships

  • Big Data Ecosystem Partners - To provide a comprehensive big data solution set to enterprises, Red Hat plans to partner with leading big data software and hardware providers to offer interoperability. Development of certified and documented reference architectures are expected to allow users to integrate and install comprehension enterprise big data solutions.
  • Enterprise Partners - Red Hat anticipates enabling the delivery of a comprehensive big data solution to its customers through leading enterprise integration partners utilizing the reference architectures developed by Red Hat and its big data ecosystem partners.

More Stories By Pat Romanski

News Desk compiles and publishes breaking news stories, press releases and latest news articles as they happen.

@ThingsExpo Stories
The Internet of Things will challenge the status quo of how IT and development organizations operate. Or will it? Certainly the fog layer of IoT requires special insights about data ontology, security and transactional integrity. But the developmental challenges are the same: People, Process and Platform and how we integrate our thinking to solve complicated problems. In his session at 19th Cloud Expo, Craig Sproule, CEO of Metavine, will demonstrate how to move beyond today's coding paradigm ...
The cloud market growth today is largely in public clouds. While there is a lot of spend in IT departments in virtualization, these aren’t yet translating into a true “cloud” experience within the enterprise. What is stopping the growth of the “private cloud” market? In his general session at 18th Cloud Expo, Nara Rajagopalan, CEO of Accelerite, explored the challenges in deploying, managing, and getting adoption for a private cloud within an enterprise. What are the key differences between wh...
SYS-CON Events has announced today that Roger Strukhoff has been named conference chair of Cloud Expo and @ThingsExpo 2016 Silicon Valley. The 19th Cloud Expo and 6th @ThingsExpo will take place on November 1-3, 2016, at the Santa Clara Convention Center in Santa Clara, CA. "The Internet of Things brings trillions of dollars of opportunity to developers and enterprise IT, no matter how you measure it," stated Roger Strukhoff. "More importantly, it leverages the power of devices and the Interne...
"We work in the area of Big Data analytics and Big Data analytics is a very crowded space - you have Hadoop, ETL, warehousing, visualization and there's a lot of effort trying to get these tools to talk to each other," explained Mukund Deshpande, head of the Analytics practice at Accelerite, in this SYS-CON.tv interview at 18th Cloud Expo, held June 7-9, 2016, at the Javits Center in New York City, NY.
The idea of comparing data in motion (at the sensor level) to data at rest (in a Big Data server warehouse) with predictive analytics in the cloud is very appealing to the industrial IoT sector. The problem Big Data vendors have, however, is access to that data in motion at the sensor location. In his session at @ThingsExpo, Scott Allen, CMO of FreeWave, discussed how as IoT is increasingly adopted by industrial markets, there is going to be an increased demand for sensor data from the outermos...
UAS, drones or unmanned aircraft, no matter what you call them — this was their week. Our news stream was flooded with updates on the newly announced rules and regulations for commercial UAS from the FAA. So, naturally we have dedicated this week’s top news round up to highlight some of our favorite UAS stories.
Internet of @ThingsExpo has announced today that Chris Matthieu has been named tech chair of Internet of @ThingsExpo 2016 Silicon Valley. The 6thInternet of @ThingsExpo will take place on November 1–3, 2016, at the Santa Clara Convention Center in Santa Clara, CA.
Basho Technologies has announced the latest release of Basho Riak TS, version 1.3. Riak TS is an enterprise-grade NoSQL database optimized for Internet of Things (IoT). The open source version enables developers to download the software for free and use it in production as well as make contributions to the code and develop applications around Riak TS. Enhancements to Riak TS make it quick, easy and cost-effective to spin up an instance to test new ideas and build IoT applications. In addition to...
CenturyLink has announced that application server solutions from GENBAND are now available as part of CenturyLink’s Networx contracts. The General Services Administration (GSA)’s Networx program includes the largest telecommunications contract vehicles ever awarded by the federal government. CenturyLink recently secured an extension through spring 2020 of its offerings available to federal government agencies via GSA’s Networx Universal and Enterprise contracts. GENBAND’s EXPERiUS™ Application...
In addition to all the benefits, IoT is also bringing new kind of customer experience challenges - cars that unlock themselves, thermostats turning houses into saunas and baby video monitors broadcasting over the internet. This list can only increase because while IoT services should be intuitive and simple to use, the delivery ecosystem is a myriad of potential problems as IoT explodes complexity. So finding a performance issue is like finding the proverbial needle in the haystack.
When people aren’t talking about VMs and containers, they’re talking about serverless architecture. Serverless is about no maintenance. It means you are not worried about low-level infrastructural and operational details. An event-driven serverless platform is a great use case for IoT. In his session at @ThingsExpo, Animesh Singh, an STSM and Lead for IBM Cloud Platform and Infrastructure, will detail how to build a distributed serverless, polyglot, microservices framework using open source tec...
Apixio Inc. has raised $19.3 million in Series D venture capital funding led by SSM Partners with participation from First Analysis, Bain Capital Ventures and Apixio’s largest angel investor. Apixio will dedicate the proceeds toward advancing and scaling products powered by its cognitive computing platform, further enabling insights for optimal patient care. The Series D funding comes as Apixio experiences strong momentum and increasing demand for its HCC Profiler solution, which mines unstruc...
The IoT is changing the way enterprises conduct business. In his session at @ThingsExpo, Eric Hoffman, Vice President at EastBanc Technologies, discussed how businesses can gain an edge over competitors by empowering consumers to take control through IoT. He cited examples such as a Washington, D.C.-based sports club that leveraged IoT and the cloud to develop a comprehensive booking system. He also highlighted how IoT can revitalize and restore outdated business models, making them profitable ...
Presidio has received the 2015 EMC Partner Services Quality Award from EMC Corporation for achieving outstanding service excellence and customer satisfaction as measured by the EMC Partner Services Quality (PSQ) program. Presidio was also honored as the 2015 EMC Americas Marketing Excellence Partner of the Year and 2015 Mid-Market East Partner of the Year. The EMC PSQ program is a project-specific survey program designed for partners with Service Partner designations to solicit customer feedbac...
IoT offers a value of almost $4 trillion to the manufacturing industry through platforms that can improve margins, optimize operations & drive high performance work teams. By using IoT technologies as a foundation, manufacturing customers are integrating worker safety with manufacturing systems, driving deep collaboration and utilizing analytics to exponentially increased per-unit margins. However, as Benoit Lheureux, the VP for Research at Gartner points out, “IoT project implementers often ...
"delaPlex is a software development company. We do team-based outsourcing development," explained Mark Rivers, COO and Co-founder of delaPlex Software, in this SYS-CON.tv interview at 18th Cloud Expo, held June 7-9, 2016, at the Javits Center in New York City, NY.
In his general session at 18th Cloud Expo, Lee Atchison, Principal Cloud Architect and Advocate at New Relic, discussed cloud as a ‘better data center’ and how it adds new capacity (faster) and improves application availability (redundancy). The cloud is a ‘Dynamic Tool for Dynamic Apps’ and resource allocation is an integral part of your application architecture, so use only the resources you need and allocate /de-allocate resources on the fly.
Connected devices and the industrial internet are growing exponentially every year with Cisco expecting 50 billion devices to be in operation by 2020. In this period of growth, location-based insights are becoming invaluable to many businesses as they adopt new connected technologies. Knowing when and where these devices connect from is critical for a number of scenarios in supply chain management, disaster management, emergency response, M2M, location marketing and more. In his session at @Th...
Machine Learning helps make complex systems more efficient. By applying advanced Machine Learning techniques such as Cognitive Fingerprinting, wind project operators can utilize these tools to learn from collected data, detect regular patterns, and optimize their own operations. In his session at 18th Cloud Expo, Stuart Gillen, Director of Business Development at SparkCognition, discussed how research has demonstrated the value of Machine Learning in delivering next generation analytics to imp...
In his keynote at 18th Cloud Expo, Andrew Keys, Co-Founder of ConsenSys Enterprise, provided an overview of the evolution of the Internet and the Database and the future of their combination – the Blockchain. Andrew Keys is Co-Founder of ConsenSys Enterprise. He comes to ConsenSys Enterprise with capital markets, technology and entrepreneurial experience. Previously, he worked for UBS investment bank in equities analysis. Later, he was responsible for the creation and distribution of life sett...