Welcome!

Eclipse Authors: Pat Romanski, Elizabeth White, Liz McMillan, David H Deans, JP Morgenthal

News Feed Item

Concurrent Enables SQL Users to Build Big Data Applications on Hadoop in Less Than 30 Seconds

Introducing Lingual, an Open Source Software Project That Delivers ANSI-Standard SQL Technology to Easily Build New and Integrate Existing Applications Onto Hadoop

SAN FRANCISCO, CA -- (Marketwire) -- 02/20/13 -- Concurrent, Inc., the enterprise Big Data application platform company, today announced Lingual, an open source project enabling fast and simple Big Data application development on Apache Hadoop™. Leveraging the power and broad platform support of the Cascading application framework, Lingual lowers the barrier to enterprise Hadoop adoption by allowing SQL users to utilize existing skills to instantly create and run Big Data applications on Hadoop, without any new training or having to think in MapReduce.

Enterprises today are rapidly adopting Hadoop to deal with growing volumes of both unstructured and semi-structured data. However, the need for Hadoop to easily integrate with existing data management systems creates a real barrier to unlocking the full potential of Big Data and Hadoop.

Enter Lingual: True SQL for Cascading and Hadoop
With Lingual, companies can now leverage existing skill sets and product investments by carrying them over to Hadoop via a standards-based technology. Analysts and developers, familiar with SQL, JDBC or traditional BI tools, can now instantly and easily create and run Big Data applications on Hadoop, while gaining significant productivity and time-to-market benefits. Lingual presents a true ANSI-standard SQL interface and is compatible with all major Hadoop distributions whether on-premise or in the cloud. Lingual use-case examples include:

  • Data analysts, scientists and developers can now simply 'cut and paste' existing ANSI SQL code from traditional data warehouses and instantly access data locked on a Hadoop cluster.

  • Developers can use a standard Java JDBC interface to create new Hadoop applications, or use any of the Cascading APIs and languages, such as Scalding and Cascalog.

  • Being ANSI-standard compliant, companies can now query and export data from Hadoop directly into traditional BI tools.

Lingual runs on Cascading, the most widely used and deployed application framework for building robust, enterprise Big Data applications on Hadoop. Recognized companies, including The Climate Corporation, eBay, Etsy, FlightCaster, iCrossing, Razorfish, Trulia, TeleNav and Twitter are using Cascading to streamline data processing, data filtering and workflow optimization for large volumes of unstructured and semi-structured data. Cascading is also at the core of popular language extensions including PyCascading (Python + Cascading), Scalding (Scala + Cascading) and Cascalog (Clojure + Cascading) -- open source projects sponsored by Twitter. Cascading has become the most reliable and repeatable way of building and deploying Big Data applications.

Supporting Quotes

"Here at Kontagent, we are very excited about the prospect of using standard SQL to provide seamless access to the billions of events that we track daily. Rather than filtering through events and exporting them to mySQL, our customer support staff and data scientists will finally be able to work with tools they already know to query the raw data directly within our Hadoop cluster through the use of Lingual and Cascading."
-Zack Shapiro, Director of Engineering, Kontagent

"Concurrent was established with the belief that there had to be a simpler path to mass Hadoop adoption. And since day one, we have worked to create solutions that make it easier for developers to build powerful and robust Big Data application, quickly and easily. With the Lingual project, we are one huge step closer to realizing our mission."
-Chris Wensel, CTO and Founder, Concurrent, Inc.

Supporting Resources

Availability and Pricing
Lingual will soon be publicly available and freely licensable under the Apache 2.0 License Agreement. To learn more about the Lingual project, visit http://cascading.org/lingual. Concurrent also offers standard and premium support subscriptions for enterprise use. To learn more about Concurrent's offerings, please visit http://concurrentinc.com.

About Concurrent, Inc.
Concurrent, Inc. is the enterprise Big Data application platform company. Founded in 2008, Concurrent simplifies Big Data application development, deployment and management on Apache Hadoop. We are the company behind Cascading, the most widely used and deployed technology for building Big Data applications with more than 75,000 user downloads a month. Enterprises including Twitter, eBay, The Climate Corporation and Etsy all rely on Concurrent's technology to drive their Big Data deployments. Concurrent is headquartered in San Francisco. Visit Concurrent online at http://concurrentinc.com.

Add to Digg Bookmark with del.icio.us Add to Newsvine

Media Contact
Danielle Salvato-Earl
Kulesa Faul for Concurrent, Inc.
(650) 340 1982
Email Contact

More Stories By Marketwired .

Copyright © 2009 Marketwired. All rights reserved. All the news releases provided by Marketwired are copyrighted. Any forms of copying other than an individual user's personal reference without express written permission is prohibited. Further distribution of these materials is strictly forbidden, including but not limited to, posting, emailing, faxing, archiving in a public database, redistributing via a computer network or in a printed form.

IoT & Smart Cities Stories
The deluge of IoT sensor data collected from connected devices and the powerful AI required to make that data actionable are giving rise to a hybrid ecosystem in which cloud, on-prem and edge processes become interweaved. Attendees will learn how emerging composable infrastructure solutions deliver the adaptive architecture needed to manage this new data reality. Machine learning algorithms can better anticipate data storms and automate resources to support surges, including fully scalable GPU-c...
Machine learning has taken residence at our cities' cores and now we can finally have "smart cities." Cities are a collection of buildings made to provide the structure and safety necessary for people to function, create and survive. Buildings are a pool of ever-changing performance data from large automated systems such as heating and cooling to the people that live and work within them. Through machine learning, buildings can optimize performance, reduce costs, and improve occupant comfort by ...
The explosion of new web/cloud/IoT-based applications and the data they generate are transforming our world right before our eyes. In this rush to adopt these new technologies, organizations are often ignoring fundamental questions concerning who owns the data and failing to ask for permission to conduct invasive surveillance of their customers. Organizations that are not transparent about how their systems gather data telemetry without offering shared data ownership risk product rejection, regu...
René Bostic is the Technical VP of the IBM Cloud Unit in North America. Enjoying her career with IBM during the modern millennial technological era, she is an expert in cloud computing, DevOps and emerging cloud technologies such as Blockchain. Her strengths and core competencies include a proven record of accomplishments in consensus building at all levels to assess, plan, and implement enterprise and cloud computing solutions. René is a member of the Society of Women Engineers (SWE) and a m...
Poor data quality and analytics drive down business value. In fact, Gartner estimated that the average financial impact of poor data quality on organizations is $9.7 million per year. But bad data is much more than a cost center. By eroding trust in information, analytics and the business decisions based on these, it is a serious impediment to digital transformation.
Digital Transformation: Preparing Cloud & IoT Security for the Age of Artificial Intelligence. As automation and artificial intelligence (AI) power solution development and delivery, many businesses need to build backend cloud capabilities. Well-poised organizations, marketing smart devices with AI and BlockChain capabilities prepare to refine compliance and regulatory capabilities in 2018. Volumes of health, financial, technical and privacy data, along with tightening compliance requirements by...
Predicting the future has never been more challenging - not because of the lack of data but because of the flood of ungoverned and risk laden information. Microsoft states that 2.5 exabytes of data are created every day. Expectations and reliance on data are being pushed to the limits, as demands around hybrid options continue to grow.
Digital Transformation and Disruption, Amazon Style - What You Can Learn. Chris Kocher is a co-founder of Grey Heron, a management and strategic marketing consulting firm. He has 25+ years in both strategic and hands-on operating experience helping executives and investors build revenues and shareholder value. He has consulted with over 130 companies on innovating with new business models, product strategies and monetization. Chris has held management positions at HP and Symantec in addition to ...
Enterprises have taken advantage of IoT to achieve important revenue and cost advantages. What is less apparent is how incumbent enterprises operating at scale have, following success with IoT, built analytic, operations management and software development capabilities - ranging from autonomous vehicles to manageable robotics installations. They have embraced these capabilities as if they were Silicon Valley startups.
As IoT continues to increase momentum, so does the associated risk. Secure Device Lifecycle Management (DLM) is ranked as one of the most important technology areas of IoT. Driving this trend is the realization that secure support for IoT devices provides companies the ability to deliver high-quality, reliable, secure offerings faster, create new revenue streams, and reduce support costs, all while building a competitive advantage in their markets. In this session, we will use customer use cases...