Cognos Insight Import Data Issue

Are you having issues with Insight not being able to import data?  It has been found that the root cause of this is IE 11.  By downgrading IE 11 we will be able to have import data work again.

Symptom:  When selecting “Import Data…” from the Get Data menu Insight will be stuck on this menu, and will not proceed to import data.

Solution:  Follow these instructions to downgrade IE11

http://www.wikihow.com/Uninstall-Internet-Explorer-11-for-Windows-7

 

Note: This issue is for Insight 10.2.1  and earlier.  The issue has been reported that it should be fixed in the next fix pack or release.

 

 

Posted in Analytics, Cognos Insight | Tagged , , , , , , | Comments Off on Cognos Insight Import Data Issue

Designing Cognos Insight

One of the most powerful parts of Cognos Insight is the ability to add different themes to your workspace.  This can cause your widgets to pop out a little more, and give a little more color.  Sometimes we want a little bit more than what we are given in the built-in themes. At this point Insight only has a few themes. All they give you is a background and a new color scheme.

I will be showing you how to make a theme that fits your needs and the needs of your business. I will be showing how to use basic themes, backgrounds, and buttons.  The place where I like to start is with your own personal logo. Continue reading

Posted in Analytics, Business, Business Analytics, Business Intelligence, Cognos, Cognos Insight, Data Integration, IBM, Productivity, Reporting | Tagged , , , , , , , | Comments Off on Designing Cognos Insight

How to Work With Dates in Cognos Insight

Dealing with dates in Cognos can sometimes be a daunting task.  Dates should be straightforward, but sometimes they take extra work to get what you want.  Your data is designed to list the date of the transaction, but maybe you only care about data on a per week, per month, or per quarter basis.  Obviously, your data isn’t arranged according to these parameters, but it is quite easy to set your dates the way you want.

Continue reading

Posted in Analytics, Business, Business Analytics, Business Intelligence, Cognos, Data Integration, IBM, OLAP, Reporting, Uncategorized | Tagged , , , , | Comments Off on How to Work With Dates in Cognos Insight

Let’s Make A Deal! Intuition, Analytics, and Answering the Right Question

I was wandering around YouTube one weekend looking for something mentally stimulating when I came across the Monty Hall problem.  For those who are not familiar, the problem is named after the host of the gameshow Let’s Make A Deal.  On the show, there was a game where the contestant would be provided 3 doors; One had a car behind it, the others had goats.  The contestant would choose the door they think had the car behind it, and then Monty would help them out by revealing one of the goats.  The contestant would then be given the chance to stay with their first choice, or switch to the other unopened door.  So what would you choose, and does it even matter?  I generated some data to simulate it, and the results will probably surprise you.

Continue reading

Posted in Business Intelligence, Cognos, Cognos Insight, Data Integration, IBM, Reporting | Tagged , , , , , | Comments Off on Let’s Make A Deal! Intuition, Analytics, and Answering the Right Question

Joining Columns In Cognos Insight

One of the benefits of Cognos insight is the ability to combine columns during the import process.  This can help us when browsing the data through explore points.  It is often beneficial to combine columns as it is easier to read, and not always necessary to sort by columns individually.

Continue reading

Posted in Business Intelligence, Cognos, IBM, MySQL, OLAP, Uncategorized | Tagged , , | Comments Off on Joining Columns In Cognos Insight

Using MySQL (or other ODBC connections) with Cognos

MySQL is quite possibly the most commonly used database today due to its free license tier making it a very low risk investment.  While it is not well geared towards the large queries of reporting and analysis, there may be a need to report directly from an existing MySQL database for any number of reasons.

The issue with MySQL in Cognos, is that it is only supported via general ODBC drivers.  The trick to remember with ODBC drivers is that even if the OS and Cognos installation are 64-bit, ODBC drivers must be 32-bit to work with Cognos.  Interestingly enough, Cognos Insight can use 64-bit drivers (probably due to it inheriting more from TM1 than Cognos proper).  Also note that ODBC drivers cannot run in Dynamic Query Mode.  This guide will step you through the process of setting up a MySQL data source specifically, but can be easily adapted to any ODBC driver.

Continue reading

Posted in Business Intelligence, Cognos, IBM | Tagged , , | 1 Comment

Raising the Bar with Mobile BI

(Enjoy a sneak peak of some dashboard screenshots as you read)

In recent years, mobile device sales have outshined those of personal computers. One survey reported that 95 percent of employees purchased a mobile device with the intent to use it for work. For many this has meant discovering the convenience of the smartphone, taking full advantage of the opportunity to integrate applications and conduct business in the palm of their hand. By 2015, 50% of devices used in business organizations will have gone mobile.

Newer and also larger than the smartphone is the tablet. These are a convenient replacement for a laptop when you’re on the go as they have foregone the traditional keyboard for an on screen version. However, they are still capable of many tasks you would find on your normal desktop or laptop.

________________________________________________________________

Example – Dashboard for Sales Executive Depicting Revenue Trends:

Image captured on IPad

 

Curious about implementation in the workforce?

  • Mobile BI has allowed retailers to make the best decisions for their businesses without the hassle of having to always be at their desk. With the tap of a finger they are able to access information on how well their marketing campaigns are doing, or review buying patterns – all while catching a cab to their 2:00 production meeting.
  • In the financial services industry, bankers are being enabled to make better decisions and analyze risk levels with greater accuracy.  Information is being integrated between silos creating ONE set of numbers increasing transparency.
  • Recent research has revealed that about 3 out of every 10 doctors are currently using IPads at work, making it easy to obtain important patient information at a moments notice.

Business Intelligence applications for tablets and cell phones are now an integral part of the mobile world. Many of the tools used by Datamensional consultants have mobile capability, and we are doing our part to make it easier than ever to keep your data close at hand. Call us for more information on optimizing your BI experience from your mobile device.

1-888-966-DATA (3282)

Additional Screenshots to convey potential for Mobile BI:

Dashboard Depicting Key Performance Indicators for the Cincinnati Zoo:

Image captured on IPad

Example of Business Analytics Delivered via Android Smartphone:

Image Captured on Android Smartphone


Posted in Business, Business Analytics, Business Intelligence, Productivity, Uncategorized | Tagged | Comments Off on Raising the Bar with Mobile BI

Big Data, What is it Exactly? Datamensional’s Take

Joseph A. di Paolantonio
Benjamin B. Goewey

Much of the current hype in Data Management & Analytics today is around the concept of Big Data, and Hadoop is at the center of the hype storm. The other two hot areas are Mobile & Elastic Cloud Computing. Cloud is central to both Big Data & Mobile implementations. This blog post will focus on Big Data, and how Datamensional has helped its customers meet this challenge with tools from Pentaho, Microsoft and IBM that work with Hadoop and other NoSQL data management systems.

In 2010 February, I suggested this approach to big data:

“Big data really isn’t about the amount of data (TB & PB & more) so much as it is about the volumetric flow and timeliness of the data streams.  It’s about how data management systems handle the various sources of data as well as the interweaving of those sources.  It means treating data management systems in the same way that we treat the Space Transportation System, as a very large, complex system.”

Many now flock to the definition of Big Data as three Vs. These Vs were debated hotly throughout 2011, on Twitter, blogs, and journals, with more Vs added. One good example can be found in R. Ray Wang’s article on Forbes:

http://www.forbes.com/sites/raywang/2012/02/27/mondays-musings-beyond-the-three-vs-of-big-data-viscosity-and-virality/

The three Vs on which everyone agrees are:

  • Volume – essentially, more data than you are accustomed to handling on your current computing platform, whether that’s Excel on your laptop, Oracle on a *nix box, or SAS on a mainframe
  • Velocity – from request a report from IT & wait a week or three, to near-real-time, [undefined, but not really] real-time and streaming data (a.k.a. Continuous Event Processing (CEP))
  • Variety – typically defined as structured (Entity Relationship Diagram (ERD) or schæma-on-write modeled data in your standard RDBMS), semi-structured(such as XML] or unstructured (with email, documents & tweets as common examples)

In the article cited above, Ray adds Viscosity and Virality (not to be confused with virility).  Viscosity can be seen as anything that impedes the interweaving or flow of data to create insights. Virality – how quickly an idea goes viral on the interwebs – or the rate at which ideas are dispersed across various Internet or Social Media sites (Twitter, YouTube, LinkedIn, Blogs, etc.).

Compare these five Vs to my definition of Big Data, given above; interweaving a heavy volumetric flow of multiple types of complex data from a variety of sources.

There are many sources of Big Data, both internal and external. To name just a few:
– Social Media
– Smartphones
– Weblogs
– Server & Network Logs
– Sensors
– The Internet of Things

Getting tons of TB to PB of data off of one data center into one cloud or vice-versa, or from one cloud to another, is logistically ridiculous. What’s in a cloud, generally stays in that cloud. Getting Analytics closer to the data is paramount for any practical application. Many vendors have recognized this, and over the past few years, the Analytic Database Management System (ADBMS), Hadoop and Cloud *aaS (* being software, infrastructure, platform, or data – as a Service) markets evolving at a pace not seen in decades, with Pentaho, SPSS, SAS, The R Statistical Language, and other analytical software being embedded in ADBMS or Cloud offerings from Teradata/Asterdata, EMC/Greenplum, HP/Vertica, SAP/Bobj/HANA/SybaseIQ, Oracle and IBM/Cognos/Netezza on the one hand, and AWS, MS Azure & Cloudera on the other.

Among others, IBM, Microsoft & Pentaho offer tools to improve Analytics out of Hadoop and other Big Data data sources. This is the most important for Datamensional. Let’s look at one customer case study, using Pentaho Hadoop Data-integration [PHD], Hadoop HDFS & Hive, and 50,000+ rows of data from one cell phone every minute. I had the honour of working with Ben & Gerrit of Datamensional on this project.

The Status Quo: Homegrown reporting & ETL solution written in Java and leveraging some open source tools and libraries to transform cell tower log files into CSV files for loading into an Oracle Datawarehouse with web-based administration & reporting.

The Business Need: Provide exploratory & in-depth Analytics capabilities to customer business analysts on a rich data set that was growing at a mind-boggling pace as Smartphone use opened new avenues for services such as location-based advertising and understanding cell phone user habits. The sample data showed 50,000 records describing a single mobile device usage over a one minute time period.

The Datamensional Solution: A Proof of Concept comparing the homegrown solution against Hadoop & Hive with the integrated Pentaho solution for Hadoop, including PHD and the pluggable architecture of the Pentaho BI Server. The PoC provided in-depth solutions in 6 areas:

  1. Installation, configuration and performance comparisons of using Pentaho Hadoop Data Integration.
  2. Demonstrating Pentaho’s Plug-in Architecture and use as a BI Platform
  3. Connecting Pentaho Analyzer to third party OLAP engines
  4. Demonstrating Pentaho Clustering & Parallel Processing capabilities
  5. Customizing the look & feel of the Pentaho User & Administration Consoles
  6. Automating the Installation Process including customizations

The Results: All points of the PoC were exceeded. While all six points are important for developing a Big Data solution or product using Pentaho, the first & fourth points, regarding PHD & Clustering are the most important for Big Data & Big Data Analytics. Other areas of the PoC showed the flexibility of the Pentaho Business Analytics platform for Reporting, OLAP,  Data Mining & Dashboards using Pentaho and other solutions, both from the Pentaho Community and as integrated by the Datamensional team.

One amazing result was that replacing the Hadoop libraries and native Hive JDBC with the Pentaho PHD versions improved response time on a simple report from so long that the customer killed the job rather than wait any longer with their homegrown solution, to less than 10 seconds with PHD. Of note, is that the PHD libraries replace the native Hadoop lib directory & files on the Hadoop name node and ALL data nodes. Clustering PDI by installing the lightweight Jetty server, Carte, on each node, parallelizes data integration and increases throughput for Hadoop & Hive. PDI provides many mechanisms to tune the performance of individual transformation steps and job entries, as well as clustering, parallelizing and partitioning for Transformations and Jobs.

The PHD libraries allow PDI and Hadoop to each bring their strengths to the data management challenges of moving, controlling, cleaning and pre-processing extreme volumetric flows of data. Using the native Hadoop libraries, a load of the sample data took over 5 hours.  This was reduced to 3 seconds using the PHD libraries and the PDI client Spoon to create transformation and orchestrate the job among the various clusters.

Additional Resources:

Wikipedia Definition

TDWI – The Three Vs of Big Data Analytics

O’Reilly Radar: What is Big Data?

Quora: What is a Good Definition of Big Data?

Datamensional Resources:

IBM on Big Data

Datamensional’s Big Data Integration Service

Pentaho Big Data Preview

Pentaho Community: Efficiency of using ETL for Hadoop vs. Code or PIG

Microsoft Case Study: MS BI and Hadoop

Posted in Big Data, Business Analytics, Data Integration, ETL, Pentaho, Pentaho Data Integration | Comments Off on Big Data, What is it Exactly? Datamensional’s Take

Xaction Basics – FTPing a file

Here is another little xaction that demonstrates the FTP capabilities of xactions. Just like the last one, make sure to save it in the workspace directory for Design Studio so that it displays properly when you try to edit it.

Open the ftp_prpt_report_w_parm.xaction file in Design Studio (Eclipse).

You will want to change the default values for each input.  Notice if you left it the way it was and ran it from PUC, you could change the default values to your FTP settings.  At first I tried to use an @ sign in the username from our Datamensional.com server, but the solution will not work with this.  It causes a problem because the output part of the .XACTION uses the @ sign to separate the username from the server name.  There may be a way to get around it through escape characters.

Go to your inputs in Design studio and change all the defaults for the following:

  • reportname
  • ftp_host
  • user
  • password
  • directory

Keep all others their default.  You will want to change the solution listed under Resources to one of your own reports as well

Save it to the solution directory in Design Studio.  In the BI Server, refresh the repository and double click on the .XACTION.  After completing successfully, you’ll see “Action Successful.” Log into the directory that you saved the FTP to, and it should be there.

Posted in BI Server, Pentaho | Comments Off on Xaction Basics – FTPing a file

Xaction Basics – Sending an Email

There are moments when working with the BI server that you will desire functionality that isn’t necessarily available through the server itself or any plugin that is currently out there.  In these situations, you will need xactions, which can be a little intimidating at first glance.  For this reason, we’ll provide some xactions that do some basic things for you to look at.  If you’d like to see a more in depth general introduction, check out this techcast by Mike Tarallo.

If you are coming from a fresh install with the sample database, you should have no problem running this .XACTION.  The only changes you need to make is to change the receiving email address, and the report being sent.  You should receive an email from the email you have set on your BI server as a default.

Make sure to put the file in the workspace when you open it through Pentaho Design Studio, otherwise it will not display any values.  The XML can still be edited this way, however.

After making your modification, save it to the solution directory from Design Studio (Eclipse) and go to PUC and refresh your repository.

Then double click on the solution.  You should see the following default message that shows the Action Sequence (.xaction) was successfully executed:

This is not all that pretty and actually could be another message.  It is simply saying that it was completed successfully and that long number is the unique ID for the document just created in the content repository.  At the end it shows what format it was.

You can execute this step using a URL into another application, another part of the suite, inside of PRD, or CDF.

Posted in BI Server, Business Intelligence, Pentaho | Comments Off on Xaction Basics – Sending an Email