Free Learning Campaign #Packt


packtApacktB Every day Packt Publishing is giving away books for free to help teach new tech skills

From 30th April, 2015 Packt Publishing has thrown open the virtual doors of its new Free Learning Library and offering its customers a daily chance to grab a fresh free eBook from its website. The publisher is encouraging people to learn new skills and try out new technologies and so every day it will be offering a different eBook from its huge list of titles free for anyone to download. The Free Learning Library will be open all year-round but each title will only be up for 24 hours, so make sure you keep checking back to get your hands on the latest book! Packt has well over 2000 titles published and the range of topics that could potentially feature is huge. From AngularJS to Zabbix, there’s going to be something to appeal to everyone – this is a great opportunity to try out a different technology or a new technique. All you’ll have to do is simply click on the day’s free eBook and it will instantly be added to your account. New customers are also encouraged to take advantage, with the offer being a brilliant chance to try out Packt’s great range of books and products – all that’s required is a Packt account. Find out more:http://bit.ly/1Kt6U6w #FreeLearning

Packt $5 eBook Bonanza campaign #packt5dollar


Packt $5 eBook Bonanza – Every title, every topic

Hi friends, the $5 eBook Bonanza is here!

Treat yourself to the eBook or Video of your choice for just $5 and get as many as you like until January 6th 2015. To get you started, we’ve put together the Top 20 Titles of 2014 for you to pick up here. But don’t forget, you can get ANY eBook or Video for $5 in this offer.

For specific tech, you can browse all our categories at the bottom of this page.
5dollar-300px_0

Book Review: Pentaho Analytics for MongoDB (February 2014)


Bookcover

PentahoBookCoverMongodb

Book review by: David Fombella Pombal (twitter: @pentaho_fan)

Book Title: Pentaho Analytics for MongoDB

Author: Bo Borland

Paperback: 146 pages

I would like to suggest this book if you want to get started with MongoDB document oriented storage engine and Pentaho Open Source BI suite.

Target Audience
This book is intended for business analysts, data architects, and developers new
to either Pentaho or MongoDB, who want to be able to deliver a complete solution
for storing, processing, and visualizing data. It’s assumed that you already have
experience in defining the data requirements needed to support business processes
and exposure to database modeling, SQL query, and reporting techniques.

Rating: 8 out of 10

 

Chapter 1, Getting Started with Pentaho and MongoDB, introduces you to the powerful combination of MongoDB and Pentaho and provides step-by-step guidance on how to install and configure both technologies and restore the sample MongoDB data provided with this book.
Chapter 2, MongoDB Database Fundamentals, expands on the topic of data modeling and explains MongoDB database concepts essential to querying MongoDB data with Pentaho.
Chapter 3, Using Pentaho Instaview, shows you how to visualize data by connecting Pentaho to MongoDB. You use Instaview with the sample MongoDB database to analyze and visualize the website clickstream data.
Chapter 4, Modifying and Enhancing Instaview Transformations, introduces Pentaho Data Integration (PDI)—the ETL tool used by Instaview to extract, load, and transform data from various data sources.
Chapter 5, Modifying and Enhancing Instaview Metadata, explores metadata by explaining dimensional modeling concepts and how to model metadata to better reflect business requirements.
Chapter 6, Pentaho Report Designer Fundamentals, teaches you the basics of Pentaho Report Designer (PRD) to build pixel-perfect reports sourced directly from MongoDB databases.
Chapter 7, Pentaho Report Designer Prompting and Charting, expands on the previous chapter by teaching you additional advanced PRD features. You can enhance your report with new queries, charts, and a prompt designed to make the report more interactive.
Chapter 8, Deploying Pentaho Analytics to the Web, is all about web-enabling your MongoDB data using Pentaho methods and web interfaces for connecting to, modeling, and analyzing our sample clickstream data in a web browser.

MongoDB Chapter8

 

Chapter Contents List:

  Chapter 1: Getting Started with Pentaho and MongoDB
MongoDB technology overview
Pentaho technology overview
Installing MongoDB
Installing MongoDB as a Windows service
Restoring the sample clickstream MongoDB database
Installing Pentaho

Chapter 2: MongoDB Database Fundamentals
MongoDB database objects
Sample clickstream database objects
MongoDB data modeling
Normalized models
Denormalized models
MongoDB query methods
Query exercise 1
Read operations
Query exercise 2
Query operators
Querying arrays

    Chapter 3: Using Pentaho Instaview
Accessing and connecting Instaview to MongoDB
Parsing and profiling a MongoDB collection
Adding a MongoDB query expression
Creating and saving an analysis view and Instaview

    Chapter 4: Modifying and Enhancing Instaview Transformations
Opening an existing Instaview
Data integration
Adding a new data source
CSV file input
Stream lookup
Creating a new analysis view from blended data

Chapter 5: Modifying and Enhancing Instaview Metadata
Model design with dimensions and measures
Open an existing Instaview
Modifying measures and dimensions
Session duration measure
Session count measure
Event count measure
Referring URL dimension
Other dimension changes
Creating a new analysis view

Chapter 6: Pentaho Report Designer Fundamentals
Pentaho Report Designer features
Data sources
Report elements
Aggregations and calculations
Formatting and output
Navigating through Pentaho Report Designer
Report workspace
The Structure tab
The Data tab
The Style and Attributes tabs
The palette
The main menu and toolbar
The tab toolbar
Interface reference
Creating a MongoDB connection and query
Adding a MongoDB data source
Adding and formatting report elements
Adding a message field to your report
Adding number-fields to your report
Adding calculated values to your report

   Chapter 7: Pentaho Report Designer Prompting and Charting
Adding additional MongoDB queries
Adding a bar chart query
Adding a pie chart query
Visualizing your data with charts
JFreeChart chart types
Subreports
Chart data collectors and properties
Creating a bar chart
Modifying bar chart properties
Creating a pie chart
Creating a report prompt
Creating a new parameter
Adding parameters to existing report queries
Creating subreport import parameters

    Chapter 8: Deploying Pentaho Analytics to the Web
Publishing a Report Designer report to the Web
Publishing the clickstream report
An introduction to the Pentaho User Console
Running and scheduling the clickstream report
Enabling your Instaview output for the Web
Copying and modifying the Instaview transformation
Using the Data Source Wizard to model your data
Creating a JDBC connection and default metadata model
Customizing the metadata model
Creating Analyzer Views and Dashboard Designer dashboards
Creating a map view in Analyzer
Creating a heat grid in Analyzer
Creating a dashboard using Dashboard Designer

Book Review: Pentaho for Big Data Analytics (November 2013)


Bookcoverhttps://www.packtpub.com/pentaho-for-big-data-analytics/book

Book review by: David Fombella Pombal (twitter: @pentaho_fan)

Book Title: Pentaho for BIg Data Analytics

Authors: Manoj R Patil, Feris Thia

Paperback: 118 pages

I would like to suggest this book if you want to get started with Pentaho Open Source BI tool together with Hadoop and Big Data.

Target Audience
If you are  a Data Scientist, a Hadoop programmer, a Big Data enthusiast, or a developer working in the Business Intelligence domain who is aware of Hadoop or the Pentaho tools and want to try out creating a solution in the Big Data space, this is your manual.

Rating: 7 out of 10

Chapter 1, The Rise of  Pentaho Analytics along with Big Data

This chapter serves as a brief summary of the Pentaho tools and its history around Business Intelligence field, weaving in stories on the rise of Big Data.

Pentaho Tools:

Server Applications

  • Business Analytics (BA) Server: Java-based BI system with a report management system and lightweight process-flow engine, HTML5-based web interface. In Community Edition , there is another substitute application called Business Intelligence (BI) Server

BA

  • Data Integration (DI) Server: Enterprise version only server for the ETL processes and Data Integration

Thin Client Tools

  • Pentaho Interactive Reporting: WYSIWYG type of design interface used to construct simple and adhoc reports on the fly without the need of having IT or programming skills. There are several CE alternatives as WAQR (Web Ad-Hoc Query Reporting) and Saiku Reporting.

PIRPentaho Interactive Reporting (EE)

saikurepSaiku Reporting (CE)

WAQRjpgWeb Ad Hoc Query Reporting

  • Pentaho Analyzer: An advanced OLAP viewer with support for drag-and-drop. It is an EE intuitive analytical visualization tool with the capability  to filter and drill down into data, stored in a Mondrian (Pentaho ROLAP engine) data source.

analyzer_territoryPentaho Analyzer

  • Pentaho Dashboard Designer (EE): Commercial plugin that allows users to create dashboards with an easy graphical interface

Design Tools

  • Schema Workbench: Graphical tool for creating ROLAP schemas for Pentaho Analysis (Mondrian).
  • Aggregation Designer: Generate pre-calculated tales  to improve the performance of Mondrian OLAP schemas with this tool.
  • Design Studio: An eclipse-based application and plugin, that eases the creation of business process flows with a special XML script to define action sequences xactions.
  • Report Designer: A banded report designing tool with a great GUI, useful to create sub-reports, charts and graphs.
  • Data Integration:  This wonderful ETL tool is also known as Kettle, and is composed by an ETL engine and GUI  that allows the user to design ETL jobs and transformations.
  • Metadata Editor: This tool is used to create business models and acts as an abstraction layer from the underlying physical database.

 

chp1Pentaho BI Suite components

Chapter 2, Setting Up the ground

In this topic we will install Pentaho BI Suite CE and Saiku OLAP plugin from Marketplace. Besides, in the chapter we learn how to administer data sources using Pentaho User Console and Pentaho Administration Console.

chp2 marketplaceMarketplace plugin

Chapter 3, Churning Big Data with Pentaho

This chapter provides a basic understanding of the Big Data ecosystem and an example to analyze data sitting on the Hadoop framework using Pentaho. At the end of this chapter, you will learn how to translate diverse data sets into meaningful data sets using Hadoop/Hive.
This chapter covers the following subjects:
• Overview of Big Data and Hadoop
• Hadoop architecture
• Big Data capabilities of Pentaho Data Integration (PDI)  Kettle
• Working with PDI and Hortonworks Data Platform, a Hadoop distribution
• Loading data from Hadoop Distributed File System (HDFS) to Hive using PDI

Hadoop ecosystemThe Hadoop ecosystem

HDFS to hive transformationHDFS to Hive transformation

Chapter 4, Pentaho Business Analytics Tools

This topics gives a quick summary of the business analytics life cycle. We will look at several applications such as Pentaho Action Sequence and Pentaho Report Designer, as well as the Community Dashboard Editor (CDE), Community Data Access (CDA) and Community Dashboard Framework (CDF) plugins and their configuration, in order to get in touch with them.

CtoolsCtools

Hive Java queryHive Java query using User Defined Java Class Step

Chapter 5, Visualization of Big Data

This chapter provides a basic understanding of visualizations and examples to analyze the patterns using various charts based on Hive data. This chapter shows us  how to create an interactive analytical dashboard that gets data from Hive. Summarizing this chapter covers the following themes:
• Evolution of data visualization and its classification
• Data source preparation
• Consumption of HDFS-based data through HiveQL
• Creation of several types of charts
• Making charts more attractive using styling

hive query chp5Hive query

DashboardStock Price Analysis Dashboard

Appendix A, Big Data Sets

Talks about data preparation with one sample from stock exchange data.

Appendix B, Hadoop Setup

Takes you through the installation and configuration of the third-party Hadoop distribution, Hortonworks Sandbox, which is used throughout the book .

http://hortonworks.com/products/hortonworks-sandbox/

Hortonworks

 

Book Review: Pentaho Data Integration Cookbook – Second Edition


Pentaho Data Integration Cookbook, Second Edition picks up where the first edition left off, by updating the recipes to the latest edition of PDI and diving into new topics such as working with Big Data and cloud sources, and more.

0674OS_ Pentaho Data Integration Cookbook (2nd edition).jpg
https://www.packtpub.com/pentaho-data-integration-cookbook-second-edition/book

Book review by: David Fombella Pombal (twitter: @pentaho_fan)

Book Title: Pentaho Data Integration Cookbook – Second Edition

Authors: Alex Meadows, Adrián Sergio Pulvirenti, María Carina Roldán

Paperback: 462 pages

I would like to suggest this useful book since it shows us how to take advantage of all the aspects of Kettle through a set of practical recipes organized to find quick solutions to our everyday needs. Although this books covers advanced topics, all recipes are explained step by step in order to help all type of readers.

Target Audience
If you are a software developer, data scientist, or anyone else looking for a tool that will help extract, transform, and load data as well as provide the tools to perform analytics and data cleansing, then this book is for you.

Rating: 9 out of 10

Chapter 1, Working with Databases – 15 recipes

This chapter shows us how to work with relational databases with Kettle.The recipes show us how to create and share database connections, perform typical database functions (select, insert, update, and delete), as well as more advanced tricks such as building and executing queries at ETL runtime. Remember that in Kettle you can connect to MySQL,Oracle, SQL Server, PostgreSQL, db2 …. and nearly all the database engines available.

Chapter 1Inserting new records when PK has to be generated based on previous values transformation

Chapter 2, Reading and Writing Files – 15 recipes

This topic not only shows us how to read and write files (csv, txt, excel …), but also how to work with semi-structured files, and read data from Amazon Web Services S3 instances.

Chapter 2Loading data into an AWS S3 Instance transformation

Chapter 3, Working with Big Data and Cloud Sources – 8 recipes

This third chapter covers how to load and read data from some of the many different NoSQL data sources (MongoDB, HBase, Hadoop …) as well as from Salesforce.com. I would like to remark the importance of this issue of the book due to the importance of Big Data techniques nowadays.

Chapter 3 Loading data into HBaseLoading data into HBase transformation

Chapter 4, Manipulating XML Structures – 10 recipes

This topic shows us how to read, write, and validate XML  files. Simple and complex XML structures are shown as well as more specialized formats such as RSS feeds. Even an HTML page is generated using XML and XSL transformations. You should read carefully this chapter if you are used to work loading,reading, updating or validating XML files.

Chapter 4Generating an HTML page using XML and XSL sources transformation

Chapter 5, File Management – 9 recipes

This chapter demonstrates how to copy, move, transfer, and encrypt files and directories. Here you will learn how to get data from remote FTP servers, zip files and encrypt files using OpenPGP standard.

Chapter 5Encrypting and decrypting files transformation

Chapter 6, Looking for Data – 8 recipes

This issue shows you how to search for information through various methods via databases, web services, files, and more. This chapter also shows you how to validate data with Kettle’s built-in validation steps. Besides, in last recipe you will learn how to validate data at runtime.

Chapter 6Validating data at runtime transformation

Chapter 7, Understanding and Optimizing Data Flows – 12 recipes

This chapter details how Kettle moves data through jobs and transformations and how to optimize data flows (Processing jobs in parallel, splitting a stream into 2 or more, comparing streams ….).

Chapter 7Run transformations in parallel job

Chapter 8, Executing and Re-using Jobs and Transformations – 9 recipes

This chapter shows us how to launch jobs and transformations in various ways through static or dynamic arguments and parameterization. Object-oriented transformations through subtransformations are also explained.Chapter 8

Moving the reusable part of a transformation to a sub-transformation (Mapping)

Chapter 9, Integrating Kettle and the Pentaho Suite – 6 recipes

This chapter works with some of the other tools in the Pentaho suite (BI Server, Report Designer) to show how combining tools provides even more capabilities and functionality for reporting, dashboards, and more. In this part of the book you will create Pentaho reports from PDI,  execute PDI transformations from BI Server and populating a dashboard with PDI.

Chapter 9Creating a Pentaho report directly from PDI transformation

Chapter 10, Getting the Most Out of Kettle – 9 recipes

This part works with some of the commonly needed features (e-mail and logging) as well as building sample data sets, and using Kettle to read meta information on jobs and transformations via files or Kettle’s database repository.

Chapter 10Programming custom functionality using Java code transformation

Chapter 11, Utilizing Visualization Tools in Kettle – 4 recipes

This chapter explains how to work with plugins and focuses on DataCleaner, AgileBI, and Instaview, an Enterprise feature that allows for fast analysis of data sources.

Chapter 11PDI Marketplace (Here you can install all plugins available)

Chapter 12, Data Analytics – 3 recipes

This part shows us how to work with the various analytical tools built into Kettle, focusing on statistics gathering steps and building datasets for Weka (Pentaho Data Mining tool), you will also read data from a SAS datafile.

Chapter 13Reading data from a sas file transformation

Appendix A, Data Structures, shows the different data structures used throughout the book.

App ASteelwheels database model structure

Appendix B, References, provides a list of books and other resources that will help you
connect with the rest of the Pentaho community and learn more about Kettle and the other
tools that are part of the Pentaho suite.

Book link:

https://www.packtpub.com/pentaho-data-integration-cookbook-second-edition/book

Book Review: Pentaho Data Integration Beginner’s Guide – Second Edition


Hello friends today I am going to review Pentaho Data Integration Beginner’s Guide – Second Edition:

5040OS.jpgFirst of all, I would like to congratulate Maria Carina a great contributor to the community pentaho I met in person in last  Pentaho Community Meeting #PCM13 in  Sintra.

Below you can check the link to purchase the book:

https://www.packtpub.com/pentaho-data-integration-beginners-guide-2e/book

Book review by: David Fombella Pombal (twitter: @pentaho_fan)

Book Title: Pentaho Data Integration Beginner’s Guide – Second Edition

Authors: María Carina Roldán

Paperback: 502 pages

I would like to recommend this book because if you are a noob in Pentaho Data Integration you will gain a lot of knowledge of this cool tool, besides if you are advanced with PDI you can use it as reference guide book.

Target Audience
This book is an excellent starting point for database administrators, data warehouse developers, or anyone who is responsible for ETL and data warehouse projects and needs to load data into them.

Rating: 9 out of 10

Although this book is oriented to PDI 4.4.0 CE version, some new features of PDI 5.0.1 CE are listed in an Appendix of the book

Kettle version

Chapter List

Chapter 1 – Getting Started with Pentaho Data Integration
In this chapter  you learn what Pentaho Data Integration is and installing the software required to start using PDI graphical designer. As an additional task MySQL DBMS server is installed.

Chapter 1Hello world transformation

Chapter 2 – Getting started with Transformations
This chapters introduces us in the basic terminology of PDI and an introduction in handling runtime errors is performed. We will also learn the simplest ways of transforming data.Chapter 2Calculating project duration transformation

Chapter 3 – Manipulating Real-World Data
Here we will learn how to get data from different sorts of files (csv, txt, xml …)  using PDI. Besides we will send data from Kettle to plain files

Chapter 3Creation of a CSV file with random values transformation

Chapter 4 – Filtering, Searching, and Performing Other Useful Operations with Data
Explains how to sort and filter data, grouping data by different criteria and looking up for data outside the main stream of data. Some data cleasing tasks are also performed in this chapter.

Chapter 4Filtering data transformation

Chapter 5 – Controlling the Flow of Data
In this very important for ETL developers chapter we will learn how to control the flow of data. In particular we will cover the following topics: Copying and distributing rows, Splitting streams based on conditions and merging streams of data.

Chapter 5Copying rows transformation

Chapter 6 – Transforming Your Data by Coding
This chapter explains how to insert code in your transformations. Specially you will learn: Inserting and testing Javascript and Java code in your transformations and Distinguishing situations where coding is the best option, from those where there are better alternatives. PDI uses the Rhino javascript engine from Mozilla https://developer.mozilla.org/en-US/docs/Rhino/Overview . For allowing Java programming inside PDI, the tool uses the Janino project libraries. Janino es a supper-small and fast embedded compiler that compiles Java code at runtime http://docs.codehaus.org/display/JANINO/Home . In summary,always remember that code in the Javascript step is interpreted, whereas the code in User Java Class is compiled. This means that a transformation that uses the UDJC step will have much better performance.

Chapter 6Transformation with java code

Chapter 7 – Transforming the Rowset
This chapter will be dedicated to learn how to convert rows to columns (denormalizing) and converting columns to rows (normalizing) . Furthermore, you will be introduced to a very important topic in data warehousing called time dimensions.

Chapter 7Denormalizing rows transformation

Chapter 8 – Working with databases
This is the firs of two chapters fully dedicated to working with databases. We will learn how to connect to a database, preview and get data from a database and insert/update/delete data from a database.

Chapter 8List of some of the many types of databases available to connect to in PDI

Chapter 9 – Performing Advanced Operations with Databases
This chapter explains different advanced operations with databases: Doing simple and complex lookups in a database. Besides an introduction in dimensional modeling and loading dimensions is included.

Chapter 9Database lookup in a transformation

Chapter 10 – Creating Basic Task Flows
So far, we have been working with data (running transformations). A PDI transformation does not run in isolation and usually is embedded in a bigger process. These processes like generating a daily report and transfer the report to a shared repository or updating a data ware house and  sending a notification by email  can be implemented by PDI jobs. In this chapter we will be introduced to jobs, executing tasks upon conditions and working with arguments and named paramenters.

Chapter 10Creating a folder transformation

Chapter 11 – Creating Advanced Transformations and Jobs
This chapter is about learning techniques for creating complex transformations and jobs (create subtransformations, implement process flows, nest jobs, iterate the execution of jobs and transformations …)

Chapter 11Execute transformation included in a job for every input row

Chapter 12 – Developing and Implementing a Simple Datamart
This chapter will cover the following: Introduction to a sales datamart based on a provided database, loading the dimensions and fact table of the sales datamart and automating what has been done.

Appendix A- Working With Repositories
PDI allows us storing our transformations and jobs under 2 different configurations: file-based and database repository. Along this book we have used file-based option, however the database repository is convenient in some situations.

Appendix B- Pan and Kitchen – LaunchingTransformations and Jobs from the Command Line

Despite having used Spoon as the tool for running jobs and transformation you may also run them from a terminal window. Pan is a cmd-line program which lets you launche the transformations designed in Spoon, both the .ktr files and from a repository. The counterpart to Pan is Kitchen, which allows you to run jobs from .kjb files and from a repository.

Appendix C-  Quick Reference – Steps and Job Entries

This appendix summarizes the purpose of  the steps and jobs entries  used in the labs throughout the book.

Appendix D-  Spoon Shortcuts

This very useful appendix includes tables summarizing  the main Spoon shortcuts.

Appendix E-  Introducing PDI 5 features

New PDI 5 features (PDI 5 is currently available now)

Book link:

https://www.packtpub.com/pentaho-data-integration-beginners-guide-2e/book

Book Review: Pentaho 5.0 Reporting by Example: Beginner’s Guide


Hello friends today I am going to review Pentaho 5.0 Reporting by Example: Beginner’s Guide  book:

Pentaho5_Libro cover

Below you can check the link to purchase it:

http://www.packtpub.com/pentaho-5-0-reporting-by-example-beginners-guide/book

Book review by: David Fombella Pombal (twitter: @pentaho_fan)

Book Title: Pentaho 5.0 Reporting by Example: Beginner’s Guide

Authors: Mariano García Mattío, Dario R. Bernabeu

Paperback: 329 pages

I like this book because if you are a noob in Pentaho Reporting you will gain a lot of knowledge of this excellent tool, besides if you are advanced with PRD you can use it as reference book.

The best of this book is that uses Pentaho 5 recent released version and shows how to publish and interacting with reports created with Report Designer 5 into the new Business Analytics server.

You should read this book because you will learn all details of this awesome tool. Maybe you have created reports using PRD but after reading this book you will improve your reporting skills.

Rating: 8 out of 10

Pentaho 5

Chapter List

Chapter 1 – What is Pentaho Report Designer?
In this chapter it is explained what is Pentaho Report Designer and some examples of typical reports are listed: Transactional Reporting, Tactical Reporting, Strategic Reporting, and Helper Reporting. Besides it is interesting to learn about the different versions of PRD along history.

Chapter 2 – Installation and Configuration
This chapters shows us how to download, install, and configure Pentaho Reporting Designer new version 5.0. Maybe it is not funny but this chapter is useful for beginners.

Chapter 3 – Start PRD and the User Interface (UI) Layout
In this chapter you will learn how to start PRD in different operating systems. Different areas of its UI are displayed too.

Pentaho5report

Chapter 4 – Instant Gratification – Creating Your First Report with PRD
Explains how to create an initial report, how to define its data sets and how to configure the report’s sections, how to add and set insertable objects and functions, and how to preview and export your report in all available formats suchs as CSV, PDF, HTML, Excel/Excel 2007. This topic is excellent to introduce non technical users to all visualization capabilities that Pentaho Report Designer includes.

Chapter 5 – Adding a Relational Data Source
Here we will create a copy of the report created in the previous chapter and modify it so that it takes its data from a relational source. We will explain what JDBC is, what a JDBC driver is, and how it is used. Using a JDBC driver you can connect to all available database systems: MySQL, Postgres, Oracle, Infobright, Vertica, Monet, Hadoop, SQL Server, PALO, Hypersonic, CouchDB…

Datasources Pentaho5

Chapter 6 – Adding Groups
In this chapter, we will use the report we created in the previous chapter and configure it to work with groups. The information included in this chapter allows us creating reports with partial group calculations and subtotals.

Chapter 7 – Adding Parameters
This chapter will be dedicated to parameters. Through the use of parameters, the final user will be able to interact with the report. A parameter is usually represented through a user interface component, which allows the selection or input of values. Pentaho Report Designer eases the task of parameter creation by providing us a simple and intuitive interface such as : Dropdowns, MultiValue Lists, Date Pickers, Text search components…

Parameters pentaho5

Chapter 8 – Using Formulas in Our Reports
In this chapter, we will talk about formulas (PRD uses OpenFormula ). We will explain how to create them and use them. As in previous chapters, we will use a practical example to guide us as we explain how to work with formulas, creating general-use formulas that we can use as an object and specific-use formulas that we can apply to our objects’ styles and attributes. By the end of the practical example, we will see how the combination of formulas and parameters opens up new horizons for the creation and personalization of reports and allows us great flexibility in design.

Chapter 9 – Adding Charts
This part of the book is dedicated to charts. We will talk about chart functions, good and bad practices of using charts, and how to create and configure our own chartsamong other topics. Report designer uses JFreeChart to create charts. By explaining each type of chart (Bar chart, Line chart, Pie chart, Area chart, Barline chart, Ring chart, Bubble chart, Scatter-plot chart, XY chart, XY line chart, XY area chart, Waterfall chart, Radar chart, XY area line chart) , we will be able to understand which chart to use for a particular need, and how to configure the chart so that its look and feel is what we want. We will also see how charts allow us to show data from different perspectives and add value to our reports.

Radar chart

Chapter 10 – Adding Subreports
This chapter is dedicated to Subreports. We will see what Subreports are and how they work and learn about their specific features. We will also talk about the relation that exists between the main Report and its Subreports. We will be able to add information to our Reports in the form of tables, charts, and so on, whether this information is connected to our data set or not. This gives us an idea of the potential that Subreports provide.

Chapter 11 – Publishing and Running Reports in Pentaho BA Server
In this chapter we talk about Pentaho, the fastest growing, most popular, and most heavily invested in Open Source Business Intelligence (OSBI) suite in recent years. We analyze its principal characteristics and the principal projects included in the suite. With Pentaho Business Analytics Business Analytics Server, we have full BI analytic power at our fingertips through a GNU General Public License (GPL). We will see how Pentaho Report Designer (PRD) and Pentaho BA Server interact, and how we can use them to run our reports from Pentaho User Console (PUC). At the end of this chapter, we will have Pentaho BA Server running and will be able to access our reports from a web browser. Pentaho can run on any compatible application servers with JEE architectures such as JBOSS, Tomcat, WebSphere, WebLogic, Oracle AS…

Chapter 12 – Making a Difference – Reports with Hyperlinks and Sparklines
In this chapter we will see how to create and configure Hyperlinks. Hyperlinks give us great flexibility in designing our reports, as they let us drill down. Using Hyperlinks, we can create a network of reports and also re-use reports. Here we will also see a very interesting object that we can add to our reports, which will present specific information graphically: the sparkline.

Hyperlinkp5

Chapter 13 –  Environment Variables, Stylesheets, and Crosstabs
This chapter is dedicated to environment variables, stylesheets, and crosstabs. We will treat each of them in detail, defining what they are about and how to implement them in Pentaho Report Designer (PRD). Using Environment Variables allows us to interact with the BA Server that is executing our reports. That is, we can obtain the username of the user who is currently logged in and his or her roles, among other information. With Style Sheets, we can configure our report so that its look and feel varies according to certain CSS Rules. These CSS Rules are easy to configure, save, and import. Finally, we will take a look at one of PRD’s new jewels, Crosstabs. Using Crosstabs, we can show OLAP-style analytic information in our reports.

Chapter 14 – PRD Reports Embedded in Web Applications
This chapter attempts at taking not only one step forward with PRD, but many. Once this chapter is finished and the concepts presented here are understood, we will be able to create a web portal that allows users to access PRD reports, send parameters to them, and select the final presentation format. We will also be able to perform every configuration needed to have the portal working as well as have it packaged so that we can transport it and deploy it in a different environment with little modification.