Any new initiative brings with it several bytes of data to start with. When we start with the goal to derive insights, evaluating the available raw data becomes the only activity for days. Fields stop making sense with relation to another if at all when we challenge it enough but often enables modeling with precision with regard to the context.

Read more: Accenture vs McKinsey Company Inside View »

 It all began with a few spreadsheets with lots of related data from respective areas, the discovery activity started with all the enthusiasm and energy similar to past activities. When we dived deeper to build the whole picture we started to struggle in putting the individual parts together. It seemed more complex than we anticipated, after adding moving and rolling aspects to current key measures. The more we focused on deriving the key metrics, the more the data fought back. 

It was then a transformative journey where narrowing down from the goal backward for a change proved helpful. Adding in the variables and factors to account for precision along the way, moving ahead and backward in timelines, until it looked well enough to be stable. Few metrics that apparently did not look convincing enough proved useful in enhancing the accuracy of our insights. 

Keep reading: TCS vs Accenture Comparison at granular level »

A data discovery journey is not without its fair share of hurdles, but it gets more exciting when we are able to create something more than what actually existed before and what we hoped to achieve. Overturning the conventional part to whole relationships and stereotypes it was an incredible satisfaction on being able to finish painting the final picture. 

The lesson that stood out the most from this exercise was that data rarely tells its story in a linear fashion. We often walk into a discovery session expecting a clean path from raw numbers to polished insights. The reality is far messier. Columns that seem irrelevant at first glance turn out to be the missing piece three weeks into the analysis. Fields that look like they hold the answer end up being noise once you account for seasonality or regional variation. The whole to part approach forced us to question our assumptions at every stage and that questioning is what ultimately made the final output trustworthy.

When working with Tableau specifically, this approach translates into a very deliberate way of building dashboards. Instead of starting with individual charts and hoping they come together into a coherent story, we began with the big picture. What is the single most important question this dashboard needs to answer? Once that was clear, every subsequent sheet and every filter was designed to peel back one more layer from that central question. The top level view gives the audience the panorama. The drill downs and filter actions give them the parts. Neither is useful without the other.

One of the practical challenges we encountered was managing the sheer volume of dimensions and measures that came out of those initial spreadsheets. In Tableau, it is tempting to drag everything onto the canvas and see what sticks. But that approach leads to cluttered visuals that confuse more than they clarify. We had to be disciplined about what made it onto the dashboard and what stayed behind as supporting analysis in separate workbooks. Not every insight deserves a spot on the final deliverable. Some insights are stepping stones that help you reach the destination but do not need to be visible to the end user.

The rolling and moving calculations that added complexity during the discovery phase turned out to be some of the most valued components in the final product. Stakeholders rarely care about a single point in time. They want to know how things are trending. They want to see whether the current month is better or worse than the trailing average. They want to spot inflection points before they become problems. Table calculations in Tableau gave us the flexibility to build these moving windows without altering the underlying data. But getting the compute context right, deciding whether a calculation should run along the table, across the pane or at the cell level, required a level of precision that only came from deeply understanding the data structure first.

Another aspect that became apparent during this project was the importance of getting the granularity right before building anything visual. There were moments where the data sat at a daily level but the insight only made sense at a weekly or monthly aggregation. Conversely, there were metrics where monthly rollups hid the very patterns we were trying to surface. The decision of what level of detail to present is not a technical one. It is a storytelling one. You choose the grain based on what resolution makes the narrative most clear to the person looking at the dashboard. Too granular and they drown in noise. Too aggregated and they miss the signal entirely.

We also learned to appreciate the value of calculated fields that serve as intermediate building blocks rather than final outputs. In complex Tableau workbooks it is common to create a calculated field that you never place on any shelf directly. It exists solely so that another calculation can reference it cleanly. These intermediate fields act as stepping stones in the logic chain. They make the workbook easier to maintain and easier for a colleague to pick up months later without having to reverse engineer a single monolithic formula. Breaking the logic into readable steps mirrors the whole to part philosophy at the formula level itself.

The collaboration aspect of this kind of work is worth mentioning as well. Data discovery done in isolation tends to produce blind spots. When we brought people from different functional areas into the review sessions, they spotted patterns and anomalies that we had been staring at for days without noticing. A finance person looking at operational data will ask different questions than an operations person would. Those cross-functional questions often led us to the most valuable insights in the entire project. Tableau made this collaboration easier because we could share interactive workbooks on Tableau Server and let people explore on their own before coming back with questions and observations.

The aesthetic dimension of the final dashboard also played a larger role than we initially expected. Early drafts were functional but visually dense. Color was used inconsistently, labels overlapped in certain filter states and the layout did not guide the eye in any particular direction. We spent a meaningful amount of time simplifying the color palette, aligning elements to a grid and ensuring that the most critical numbers occupied the most prominent positions on the screen. A well designed dashboard should feel like a newspaper front page. The headline grabs you first, the subheadings orient you and the details are there when you choose to read further. That hierarchy does not happen by accident. It requires the same whole to part thinking applied to the visual design as was applied to the data itself.

Looking back at the full arc of this project, the biggest takeaway is that the struggle in the middle is not a sign that something is going wrong. It is the process working as intended. When the data fights back, when the pieces refuse to fit, that friction is where the real understanding gets built. The temptation to shortcut past the messy middle phase is strong especially when deadlines are tight. But every time we resisted that temptation and stayed with the discomfort a little longer, the output that came out the other side was materially better.

The whole to part panorama is not just a methodology for a single project. It is a mindset that carries forward into every subsequent analysis. Once you train yourself to start from the destination and work backward, to hold the full picture in your head while you tinker with the individual components, the quality of your work changes permanently. Data stops being a chore and starts being a puzzle where every piece has a place even if it takes a while to find it.