Skip to main content

5 posts tagged with "data-analysis"

View All Tags

Our bet on exploring Data without code

· 5 min read
Vincenzo Manto
Founder @ Datastripes

Our unspoken truth about Data Analysis

Most data tools weren’t built for you. They were designed for analysts, engineers, and data scientists—people who think in code, not questions. For everyone else, it’s like trying to do surgery with boxing gloves.

You’re often handed a beautiful dashboard, but when you want to dig deeper, you're stuck. You’re expected to extract insight without the means to explore freely. The data is there, but it’s locked behind layers of filters, dropdowns, and queries that don’t reflect how you actually think.

And when time matters, that disconnect hurts. You're left waiting on others, or worse—making decisions based on gut feeling because the tool couldn’t keep up with your brain.


Why tools don’t usually work for exploration

Let’s be honest: spreadsheets are great until they’re not. They let you poke around numbers quickly, but they collapse under the weight of complexity. One mistake in a formula and suddenly your whole logic falls apart. And if you ever tried to repeat an analysis next month, you’ll realize you have no idea how you got those numbers in the first place.

Business intelligence platforms, on the other hand, offer flashy dashboards, but they usually answer yesterday’s questions. They’re built for static reporting, not curiosity. You don’t follow your thought process—they force you into predefined views. It's like being told the ending of a movie when you really just wanted to explore the story.

Even modern data notebooks, while powerful, assume a technical background. If you don’t know how to write Python or SQL, you’re simply not invited to the party. Some tools try to bridge this gap with “low-code,” but the learning curve is still there, and the risk of breaking something often outweighs the benefits.


How real exploration works

When you’re digging into a problem, your brain works like this: you ask a question, you slice the data, you try something new, you see what changes. Then you go back. You test again. You don’t think in joins and queries. You think in steps. In stories. In flow.

So why don’t our tools work like that?

What we need is a space where data feels alive. Where your thought process is visible. Where one step leads to another, and each transformation is easy to follow. Imagine sketching your reasoning out as a path—from raw data, through filters and calculations, to insight. And being able to see it all evolve, right there on screen.


The hidden cost of waiting

Every time you need someone else to run a query, every time you get stuck fiddling with filters you don’t understand, every time you copy-paste data from one tool to another—you’re burning time and losing clarity. Not just operationally, but mentally. That friction erodes your momentum, and momentum is everything when you’re trying to make sense of complex things.

Data should feel like a conversation. Instead, it often feels like a form submission.


Rethinking the interface

What if instead of rows and dashboards, you worked with a visual canvas? Not to make things “pretty,” but to actually see how your analysis unfolds. You drag a filter into place, and the data updates instantly. You group something, and the insight appears. You layer logic like you would post-it notes, refining as you go. Each step can be traced, adjusted, undone, branched off. It’s not about making data pretty—it’s about making your thinking visible.

And sharing? That’s not exporting a chart to PDF. It’s handing someone your flow—your train of thought—so they can walk through it too.


Can this all happen in the browser?

It sounds too good to be true. But browser technology has quietly leapt forward. With WebAssembly, WebGPU, and new APIs, it's now possible to build serious, high-performance tools that run entirely on your machine, inside your browser. No servers, no syncing, no privacy concerns. You own the data. The browser does the heavy lifting.

This means real-time visualizations. It means processing large datasets client-side. It means no installs, no logins, no barriers between you and your questions.


Our bet

This is the challenge we took on. What if exploring data was like sketching ideas on a whiteboard? What if the interface wasn’t a table or a chart, but a flow—a chain of thoughts you could see and evolve?

We built it. It’s called Datastripes.

A visual, no-code data engine that runs entirely in the browser. It’s fast, local-first, and built for people who think visually but don’t write code. You can try the interactive demo on the homepage, no signup needed. And if it speaks to you, join the waitlist—we’re opening it up soon.

We’re betting this is how data tools should feel: live, intuitive, and actually fun to use. We hope you’ll bet with us.

Forecasting and Clustering in Google Colab

· 5 min read
Vincenzo Manto
Founder @ Datastripes
Alessia Bogoni
Chief Data Analyist @ Datastripes

Data analysis often involves multiple steps — cleaning, exploring, visualizing, modeling. Two common and powerful techniques are forecasting (predicting future trends) and clustering (grouping similar data points).

In this post, we’ll show how to do both using Google Colab, walk through the code, and highlight the complexity involved — then reveal how Datastripes can simplify this to just a couple of visual nodes, no code required.


Time Series Forecasting with Prophet in Colab

Suppose you have daily sales data, and you want to forecast the next 30 days. Prophet, a tool developed by Facebook, is great for this.

The Data

Imagine a CSV like this:

dsy
2024-01-01200
2024-01-02220
2024-01-03215
......

Where ds is the date and y is the sales.

Step-by-step Code Walkthrough

# Install Prophet - this runs only once in the Colab environment
!pip install prophet

This command installs Prophet in the Colab environment. It might take a minute.

import pandas as pd
from prophet import Prophet
import matplotlib.pyplot as plt

Here we import the necessary libraries:

  • pandas for data handling
  • Prophet for forecasting
  • matplotlib for plotting
# Load your sales data CSV into a DataFrame
df = pd.read_csv('sales.csv')

You’ll need to upload your sales.csv file to Colab or provide a link.

# Take a peek at your data to ensure it loaded correctly
print(df.head())

Always check your data early! Look for correct date formats, missing values, or typos.

# Initialize the Prophet model
model = Prophet()

This creates the Prophet model with default parameters. You can customize it later.

# Fit the model on your data
model.fit(df)

This is where the magic happens — Prophet learns the patterns from your historical data.

# Create a DataFrame with future dates to forecast
future = model.make_future_dataframe(periods=30)
print(future.tail())

make_future_dataframe adds 30 extra days beyond your data so the model can predict future values.

# Use the model to predict future sales
forecast = model.predict(future)

forecast now contains predicted values (yhat) and confidence intervals (yhat_lower and yhat_upper).

# Visualize the forecast
model.plot(forecast)
plt.title('Sales Forecast')
plt.show()

You get a clear graph showing past data, predicted future, and uncertainty.

Tips for Better Forecasts

  • Ensure your dates (ds) are in datetime format.
  • Check for missing or outlier data points before fitting.
  • Tune Prophet’s parameters like seasonality or holidays for your context.

Clustering Customers Using KMeans in Colab

Now, let’s say you want to segment customers based on income and spending behavior.

The Data

A CSV with columns:

CustomerIDAnnual Income (k$)Spending Score (1-100)
11539
21681
3176
.........

Step-by-step Code Walkthrough

from sklearn.cluster import KMeans
import matplotlib.pyplot as plt
import pandas as pd

We import KMeans for clustering, matplotlib for plotting, and pandas to load data.

# Load the customer data CSV
df = pd.read_csv('customers.csv')
print(df.head())

Always check the data to understand its shape and content.

# Select the two features to cluster on
X = df[['Annual Income (k$)', 'Spending Score (1-100)']]

These columns will form a 2D space for clustering.

# Initialize KMeans with 3 clusters
kmeans = KMeans(n_clusters=3, random_state=42)

Choosing number of clusters is a key step. Here we pick 3 for illustration.

# Fit the model and predict cluster assignments
kmeans.fit(X)
df['Cluster'] = kmeans.labels_

Each customer gets assigned a cluster label (0,1,2).

# Plot clusters with colors
plt.scatter(df['Annual Income (k$)'], df['Spending Score (1-100)'], c=df['Cluster'], cmap='viridis')
plt.xlabel('Annual Income (k$)')
plt.ylabel('Spending Score (1-100)')
plt.title('Customer Segmentation')
plt.show()

The scatter plot shows customers grouped by clusters in different colors.

Tips for Better Clustering

  • Normalize or scale features if they have different units.
  • Experiment with cluster counts and validate with metrics like silhouette score.
  • Visualize results to make business sense of clusters.

Why Is This Hard for Most People?

If you’re not a coder, these steps look intimidating: installing packages, writing code, understanding APIs, and debugging errors.

Even for tech-savvy folks, repeating these steps every time the data updates is tedious.

It takes time away from what really matters: interpreting results and making decisions.


How Datastripes Makes This Effortless

With Datastripes, you don’t need to write or understand code:

  • Upload your data.
  • Drag a "Forecast" node and configure date and value columns.
  • Drag a "Cluster" node, pick features, and watch clusters appear.
  • Everything updates live and visually, directly in your browser.
  • No installs, no scripts, no errors.

Datastripes is built to turn these complex workflows into intuitive flows — freeing you to focus on insight, not syntax.

Try the live demo at datastripes.com and see how forecasting and clustering go from tens of lines of code to just two nodes.


When data analysis becomes simple, you can explore more, decide faster, and actually enjoy the process.

How to use Power BI and Datastripes for data analysis

· 6 min read
Vincenzo Manto
Founder @ Datastripes

If you’re diving into data analytics, you’ve probably heard of Power BI — Microsoft’s powerful and widely used tool. But now there’s Datastripes, a fresh platform focused on making data work simple and visual, no coding needed. Let’s break down how these two stack up, so you can decide which one fits your style and needs best.

Why Datastripes Might Win the Data Race

· 5 min read
Alessia Bogoni
Chief Data Analyist @ Datastripes

In the world of data tools, it’s easy to get overwhelmed. So many platforms promise powerful analytics, dashboards, or integrations — but which one really gets you? Which one keeps things simple without sacrificing muscle?

Let’s cut through the noise and see why Datastripes stands out from the crowd — and why it might just be your new best data buddy.

The magic of Datastripes — Easy Peasy Data Squeezy!

· 8 min read
Vincenzo Manto
Founder @ Datastripes
Alessia Bogoni
Chief Data Analyist @ Datastripes

Welcome to Datastripes, the freshest, most flexible data workspace designed for anyone and everyone who wants to master their data — without headaches, without fuss, and with a whole lot of fun! Whether you’re a data newbie, a savvy analyst, or a seasoned pro, Datastripes turns your complex workflows into a smooth, flowing adventure. Think of it like LEGO blocks for data: snap together powerful tools, build workflows, and watch insights come alive — all with zero coding stress.

At the heart of Datastripes lies a rich catalog of nodes — tiny engines of magic that fetch, transform, visualize, compute, and export data — each designed with simplicity, flexibility, and fun in mind.