What is Apache Hive? Hive provides a mechanism to query, create and manage large datasets that are stored on Hadoop, using SQL like statements. It also enables adding a structure to existing data that resides on HDFS. In this post I’ll describe a practical approach on how to ingest data into Hive, with the SnapLogic Elastic Integration Platform, without the need to write code.
SnapLogic CTO James Markarian recently appeared as a guest on DisrupTV, a weekly live-interview web-series produced by analyst firm Constellation Research and hosted by R “Ray” Wang and Vala Afshar. The trio discussed a variety of enterprise topics including modern data management, data lake strategy considerations and big data analytics.
SnapLogic co-founder and CEO Gaurav Dhillon sat down recently with Scott Kupor, managing partner at Andreessen Horowitz, for a wide-ranging podcast discussion of all-things-data.
The two discussed how the data management landscape has changed in recent years, the rise of advanced analytics, the move from data warehouses to data lakes, and other changes which are enabling organizations to “take back their enterprise.”
SnapLogic was in New York this week for Strata + Hadoop World NYC, and our CTO James Markarian took the opportunity to sit down with Dave Vellante and George Gilbert, hosts of theCUBE, for a wide-ranging discussion on the shifting big data landscape.
SnapLogic announced the availability of new pre-built intelligent connectors – called Snaps – for Microsoft Azure Data Lake Store. The new Snaps provide fast, self-service data ingestion and transformation from virtually any source – whether on-premises, in the cloud or in hybrid environments – to Microsoft’s highly-scalable, cloud-based repository for big data analytics workloads. This latest integration between SnapLogic and Microsoft Azure helps enterprise customers gain new insights and unlock business value from their cloud-based big data initiatives.
Next week our team of integration experts will be in New York for Strata + Hadoop World to demonstrate how our big data integration platform as a service (iPaaS) allows customers to quickly ingest, prepare and deliver data to other sources within their IT ecosystems. We are also hosting a networking event for big data game-changers on demystifying data lakes, Hadoop and hybrid architecture. Learn more here.
In this episode of the SnapTalk podcast series, enterprise architect Ravi Dharnikota talks with Rakesh Raghavan, Director of Snap Engineering at SnapLogic. Rakesh comes to SnapLogic having designed, developed and managed data lakes for several leading online retailers and consumer-facing websites. He has successfully navigated enterprise data lakes using open source tools and manual techniques, and in this episode shares his first-hand experiences.
Ravi and Rakesh discuss the pitfalls of jumping into a data lake without a clear architecture, the challenges of supporting both traditional reporting and ad hoc data exploration use cases in the same environment, and the often-overlooked, often manual data engineering tasks involved in data lake implementation.
Subscribe to the series: https://soundcloud.com/snaplogic/sets/snaptalk