AtScale to add BI on Hadoop

AtScale Intelligence Platform 3.0 designs to give business users the capability to interactively query Hadoop data without the requirement of extensive preprocessing or movement of the cluster.

David Mariani, while serving as vice president of Engineering in the Advertiser Analytics unit at Yahoo and later as vice president of Development in the User Data & Analytics unit, with his team is responsible for running data pipelines, connecting business users with the company’s ever-growing store of data.

He states ,”We had this whole notion of never throwing data away, That was a golden rule at Yahoo!. We just about broke the bank with NetApp and EMC. That’s why we invented Hadoop.”

“The benefit of OLAP was our users [at Yahoo!] could self-serve,” Mariani states. “We generated $50 million in lift from the display advertising business every year because of Analysis Services. We wanted an OLAP interface without the OLAP baggage. Give me Hadoop, with scale-out architecture but an OLAP interface.”

With AtScale, which came out of stealth earlier this year, Mariani aims to relieve some of the pains he faced building what may have been the world’s largest OLAP cube for business intelligence (BI) at Yahoo. AtScale enables business users to use any BI tool to work with data in Hadoop by creating a virtual cube that essentially turns Hadoop into a high-performance OLAP server.
The Tableau Conference in Las Vegas, AtScale took the wraps off the newest version of its platform: the AtScale Intelligence Platform 3.0. The new version presents a number of innovations around enterprise security and scalability. In addition, it features integration with Tableau Server and Tableau Online. Maybe most importantly, it adds AtScale’s new patent-pending technology: Adaptive Cache.

Richard Langlois, director of Enterprise Data Management at Yellow Pages (Canada) told in a statement Monday morning, “Adaptive Cache is possibly one of the most meaningful breakthroughs in this space, We put the AtScale Adaptive Cache technology through a test on 38 billion rows of data. The results were beyond expectations.”

Hadoop is extensively flexible and scalable, but it won’t able to support the kind of interactive query performance that business users expect from their BI tools. To solve the issue, organizations believed on data indexing, transformation or data movement methodologies that are typically complex and time-intensive.

Adaptive Cache is an engine that leverages ongoing learning, it views user query patterns as users access the AtScale virtual cube and then caches the data that will allow the query to run faster the next time. Data change detection identifies when new data are available and generates appropriate cache components, adding a new piece into the existing cache.

Josh Klahr, vice president of Product at AtScale stated, “As the cube is used and accessed, we get smarter and smarter about delivering interactive components, the end result is the sort of interactive experience BI users expect without the need for extensive preprocessing or movement off the cluster”. Klahr further adds extensive use of the training mode can get the cube up to speed within the first few days of implementation.

Mariani states, “The industry is going through a massive transformation and enterprises need a solution that allows business users to access any data from anywhere, Enabling this level of accessibility requires sophistication and a solid and secure platform. With this release, AtScale becomes the BI on Hadoop vendor with the most enterprise-ready capabilities.”

Source : http://www.insightssuccess.com/