Skip to content
No results
  • Home
  • Articles & Blog
  • Videos
  • Training
  • Contact us
The Data Signal
  • Home
  • Articles & Blog
  • Videos
  • Training
  • Contact us
Login
The Data Signal

The Data Signal

The Data Signal
I built a Tetris-style game called TetriPulse using Codex.

But this video isn’t about “look what AI can do.”

It’s about what it actually takes to turn AI output into working software.

In this breakdown, I walk through:

• How I structured the project
• The architecture decisions behind the game loop
• How I wrote and refined prompts
• Why context management matters more than people realise
• The iteration cycles that shaped the final product

AI can generate code.

But building software still requires:

– Design thinking
– Systems awareness
– Trade-off decisions
– Continuous refinement

This is a transparent look at what the real workflow looks like when you combine AI with engineering discipline.

If you’re building with AI, or thinking about it — this is the process behind the scenes.

🔗 Repo

TetriPulse on GitHub:
https://github.com/david-ikenna-ezekiel/tetripulse
Inside My AI Game Build: Architecture, Prompts & Iterations
Ever opened your Downloads folder and realised it’s complete chaos?

In this video, I show how I used Codex CLI to automatically organise my messy downloads folder into structured categories — separating files to keep, files to remove, and grouping them by file type.

This wasn’t a polished production script… just a fun real-world experiment showing how AI can automate everyday digital clutter.

By the end of this video, you’ll see how developer AI tools like Codex can go beyond coding and help automate real life workflows.

🚀 What You’ll Learn

- How Codex CLI works in practice
- Using AI to analyse and categorise files
- Automating folder organisation
- Structuring files into meaningful categories
- Practical AI automation ideas you can replicate
I Let Codex CLI Clean My Messy Downloads Folder…Here’s What Happened
Want to become a data analyst in 2026 but don’t know where to start?

This video breaks down the only roadmap you need — step by step — starting from absolute zero. No degree required. No advanced math. Just the right order of skills that actually get people hired.

Stuff inside:

- Where beginners should actually start

- Which skills matter most in 2026

- Why learning tools in the wrong order slows you down

- How to build job-ready data analyst skills from scratch

If you’re serious about breaking into data analytics, save this video and follow for the next steps.
The Only Roadmap You Need to Become a Data Analyst in 2026!! #dataanalytics #datascience #AI
ETL vs ELT is one of those data concepts that sounds simple — until you actually try to explain it.

Most explanations focus on the acronyms:
Extract, Transform, Load vs Extract, Load, Transform.

But the real difference isn’t the order of the letters.

In this video, we break down ETL vs ELT by looking at what actually matters in modern data pipelines:

- Where transformations really happen

- Why ETL dominated for decades

- What changed with cloud data warehouses

- Why ELT fits modern analytics workflows

- And when ETL still makes more sense today

Instead of definitions and buzzwords, this explanation focuses on mindset, infrastructure, and real-world use cases — so the difference finally clicks.

If you’ve ever heard ETL and ELT discussed in meetings and felt like you kind of understood it, but not fully, this video is for you.
ETL vs ELT: The Real Difference in Modern Data Pipelines
In this video, I walk through how to deploy a simple website on Google Cloud Platform using a Linux virtual machine and Apache HTTP Server.

We start by creating a Linux VM in Compute Engine, logging into it via SSH, installing Apache, and finally serving a custom HTML page to the public internet. This is a hands-on, beginner-friendly walkthrough designed to help you understand how cloud virtual machines and web servers actually work together.

What you’ll learn:

- How to create a Linux VM in Google Compute Engine
- How to SSH into your VM
- How to install and run Apache HTTP Server
- How to create and serve a custom HTML website
- How to access your site via a public IP

Perfect if you’re new to GCP, cloud infrastructure, or want to understand the fundamentals behind hosting websites on virtual machines.
Deploy Your First Website on Google Cloud in 5 Minutes (Linux VM + Apache)
Want to use Google Drive like a fully functional cloud storage system for your Databricks projects?
In this step-by-step tutorial, I’ll show you the exact trick to connect Databricks to Google Drive using OAuth + PyDrive2, download any file type, and even auto-convert Google Sheets → CSV inside your notebook.

This method turns Google Drive into a lightweight, flexible data lake — perfect for analytics, machine learning, and data engineering workflows.

What You’ll Learn

✔️ How to enable Google Drive API
✔️ How to create OAuth credentials in Google Cloud
✔️ How to authenticate Databricks using PyDrive2
✔️ How to download files directly from Drive (CSV, XLSX, PDFs, ZIPs, etc.)
✔️ How to export Google Sheets automatically to CSV
✔️ How to load Drive data into Spark
✔️ How to build a reusable Drive download function

📁 Notebook & Code

Download the full Databricks notebook here 👇
🔗 GitHub Repository: [https://github.com/david-ikenna-ezekiel/thedatasignal/tree/main/google-drive-databricks]


🛠️ Technologies Used

- Databricks
- PyDrive2
- Google Drive API
- Google Cloud Platform
- Spark
- Python

👍 Like, Share & Subscribe

If this tutorial helped you, consider subscribing for more Databricks, Big Data, and automation content.
This Trick Turns Google Drive Into a Data Lake for Databricks
Struggling to wrap your head around SQL Window Functions? You’re not alone, they’re one of those things that sound complex until someone explains them the right way.

In this video, I’ll break down SQL Window Functions from scratch, in a way that finally makes sense.

We’ll cover:
✅ What window functions are and why they exist
✅ How OVER() and PARTITION BY actually work
✅ The difference between aggregates and window functions
✅ Real examples using Google BigQuery
✅ And how to calculate running totals, rankings, and moving averages — step by step.

By the end, you’ll not only understand window functions, you’ll actually use them confidently in your own SQL queries.

💡 What You’ll Learn

- How to calculate running totals and rolling averages with ROWS BETWEEN
- Ranking and ranking differences: ROW_NUMBER(), RANK(), and DENSE_RANK()
- Comparing each row to group averages using PARTITION BY
- Why window functions beat subqueries for analytics
- How BigQuery handles partitions and frames behind the scenes

🧠 Who This Video Is For

- Data analysts & data scientists learning advanced SQL
- Developers who use BigQuery, PostgreSQL, or Snowflake
- Anyone who’s tired of copy-pasting SQL code without really understanding it

⏰ Timestamps

00:00 – Intro
01:03 – Why Window Functions Exist
02:37 – Why Windows Function is Important
03:35 – The Magic of Over()
04:40 – Demo Practice

🙌 Support the Channel
If this video helped you finally “get it,” please like, subscribe, and share it with a friend who’s learning SQL too!

Let’s make data make sense — one query at a time 💙
Stop Struggling with SQL Window Functions — Watch This First!
Ever wondered what data engineers actually do? 🤔 It’s not just writing SQL or building pipelines all day. In this short video, I break down the 5 real things data engineers do, in plain English, with simple, real-world examples.

Whether you’re just starting out in tech, curious about data careers, or trying to understand what your company’s data team really does—this is for you.

You’ll learn how data engineers:
1️⃣ Build the data plumbing that keeps information flowing
2️⃣ Clean up messy, chaotic data so others can trust it
3️⃣ Organize everything in data warehouses like Snowflake or BigQuery
4️⃣ Make systems fast, reliable, and scalable
5️⃣ Help analysts, data scientists, and teams actually use the data
What do data engineers do all day? #ai #dataengineering #job
5 rookie mistakes data engineers still make—fix these and earn trust fast. #DataEngineer #Airflow #dbt #Prefect #Dagster #GreatExpectations
5 Habits Every Data Engineer Must Drop -- they're career killers #dataengineering #programming #ai
In this video, we explore how serverless compute works in data engineering and compare it with traditional clusters. If you’re learning modern data platforms like BigQuery, Snowflake, or Databricks, understanding the difference between clusters and serverless is essential.

You’ll learn:

- What clusters are and how they work in data platforms
- What serverless compute means (and why it’s different)
- Pros and cons of clusters vs serverless
- How BigQuery, Snowflake, and Databricks use these models today

By the end of this video, you’ll have a clear understanding of why serverless computing is so powerful for modern data engineering and how it changes the way we process data.

🚀 This is part of my beginner-friendly series on data platforms and cloud data engineering. Subscribe for more videos on BigQuery, Snowflake, Databricks, and other modern data tools!
How does Serverless Compute Work in Data Engineering? Clusters vs Serverless Explained
Understanding compute and storage is one of the most important first steps in becoming a data engineer. In this video, I break down what compute is, what storage is, and why separating them matters in modern data platforms.

Whether you’re just starting out in cloud data engineering or brushing up on the fundamentals, this video explains these concepts with simple analogies and practical examples.

What you’ll learn in this video:

- What storage means in data systems (where data lives)
- What compute means (the power to process data)
- Why older systems tightly coupled compute + storage
- How modern platforms (BigQuery, Snowflake, Databricks) separate them for scalability and cost efficiency

🚀 This is part of my beginner-friendly series on data platforms — subscribe if you want to keep learning about cloud data engineering, serverless computing, and big data tools.
How Modern Data Platforms Separate Compute & Storage | Data Engineering Basics
In this video, I put ChatGPT Agents to the test with two real-world demos plus an overview of how they work. You’ll see exactly how I went from a simple prompt to full automation — no coding from scratch!

What’s inside:
00:00 – Introduction & Overview of ChatGPT Agents
06:54 – Demo 1: Create a polished presentation directly from a Notion page
11:47 – Demo 2: Process messy, unstructured data from Azure Blob Storage and load it into Airtable (ETL pipeline)
25:41 – Wrap up & key takeaways

Why this matters:
✅ Automate repetitive workflows
✅ Clean and structure messy data
✅ Connect multiple tools without complex coding
✅ Speed up your data and content processes with AI

Tools used in this video:

- ChatGPT Agents
- Notion
- PowerPoint
- Azure Blob Storage
- Airtable Web API

📌 Link to Resources:

All resources, including prompts can be found here: https://cultured-weeder-e32.notion.site/Chat-GPT-Agent-Mode-247f75d4b06780208c51feb7848a6f2c?source=copy_link 

If you found this helpful, subscribe for more AI-powered workflow demos, automation tips, and data engineering tricks!

#ChatGPTAgents #ETL #Automation #Airtable #Notion #AzureBlob #DataEngineering #AI
BEST DEMO on ChatGPT AGENTS: Build Slides from Notion, Clean Data, Load to Airtable
Subscribe

You may Have Missed

SQL Command Essentials: CREATE, INSERT, DELETE, UPDATE
How to Provision the Medallion Architecture on Azure ADLS using Terraform
Mastering SQL Aggregates and Groups for Powerful Data Analysis
Image Source: https://pin.it/5Atp5uPHZ
SQL Joins & Advanced Queries
About The Data Signal

The Data Signal is a data analytics site built to teach young data analysts all the rudiments in the field.

Site Navigation

Home | Articles & Blog | Videos | Trainings | Consulting | Contact us

Copyright © 2026 - The Data Signal. All rights reserved.

Download Guide for Free

Document will be sent to your email address.

Email Collection Form