Skip to content
View TubaSid's full-sized avatar
👽
Focusing
👽
Focusing

Highlights

  • Pro

Block or report TubaSid

Block user

Prevent this user from interacting with your repositories and sending you notifications. Learn more about blocking users.

You must be logged in to block users.

Maximum 250 characters. Please don't include any personal information such as legal names or email addresses. Markdown supported. This note will be visible to only you.
Report abuse

Contact GitHub support about this user’s behavior. Learn more about reporting abuse.

Report abuse
TubaSid/README.md

Tuba Siddiqui

where data meets intent and systems learn to think

typing

The Journey

Most people use frameworks. I'm learning to build them.
Most people read papers. I'm implementing them.
Most people know what works. I'm discovering why.

This is a self learning journey into the architecture of intelligence itself. From the mathematics of attention to the pragmatics of production. From transformer implementations in NumPy to RAG systems that ship.

Credentials: MSc Data Science @ Sapienza University of Rome
Experience: Production systems @ Fastweb (Italy) & HBL (Pakistan)
Philosophy: Every abstraction hides a lesson. Every framework makes assumptions. Know both.


What I Build

Khayalmemory as conversation

An AI companion that remembers. Integrated with WhatsApp for daily reflections, emotional check ins, habit tracking. Not just responding but understanding continuity.

Stack: FastAPI • WhatsApp API • Vector Memory Read how it thinks

Feedback Collectorsignal from noise

Automated intelligence that captures scattered feedback, finds patterns, generates summaries. The system that thinks about what users think.

Stack: n8n • Gemini • Workflow Orchestration

Read how it thinks


Current Learning Arc

Currently inside: "Attention Is All You Need" (Vaswani et al.)
Implementing every component from scratch. No tutorials. Just math, paper, and NumPy.

Also reading:

  • GPT-3 paper (architecture evolution)
  • RAG paper (Lewis et al., 2020)
  • LLaMA technical report (understanding what changed)

This month's experiments:

  • ✓ Self-attention mechanism (50 lines, pure NumPy)
  • ⟳ Building RAG without frameworks (5 different approaches)
  • → Comparing LangChain vs pure implementations

Recent Thoughts

Latest writing: The Feedback Collector That Thinks
On building systems that understand user intent without human intervention

What broke recently: Discovered my RAG was confidently hallucinating because chunking strategy was too naive. Fixed with semantic splitting and reranking. Latency increased 40% but accuracy improved 60%. Worth it.

Current question: When do frameworks actually help vs just hide complexity? Building 5 RAG implementations to answer with data.


Learning Metrics

Papers implemented from scratch        → 3
RAG systems built this month           → 2 (working on 3 more)
Lines of attention code written        → ~847
Failed experiments documented          → 12
Production tradeoffs measured          → cost • latency • accuracy

What I've Studied

Fetal Brain Structure SegmentationMaster's Thesis
Attention U-Net meets Grad-CAM. Deep learning for medical anomaly detection in ultrasound imagery.
→ Repository

VisionRoadBachelor's Engineering Thesis
YOLOv3 and Raspberry Pi teaching traffic lights to adapt.
→ Project

Voice Recognition System
MATLAB, DSP, and the mathematics of identifying speakers.
→ Details

Address Data Cleaning Tool
Fuzzy matching and postal validation. Making messy data make sense.


The Craft

Languages: Python • R • MATLAB
Learning: PyTorch • TensorFlow • NumPy • FAISS
Building: FastAPI • Streamlit • Docker
Deploying: AWS • n8n workflows
Exploring: OpenAI • Anthropic • Groq
Analyzing: PowerBI • OpenCV


Collaboration

Open to discussing:

  • RAG optimization strategies (chunking, reranking, hybrid search)
  • Production LLM systems (cost vs latency vs accuracy tradeoffs)
  • Building AI systems without heavy frameworks
  • Technical writing about deep learning journeys

Would love to collaborate on:

  • Projects that require understanding systems from first principles
  • Building tools for other learners diving deep into AI
  • Writing about the gap between research papers and production

Best way to reach me: Email or LinkedIn for thoughtful technical discussions


un'armonia d'immaginazione

AI is an orchestra. Data is its melody. Architecture is the score we compose.


STUDY
papers deeply
code completely

BUILD
systems carefully
production ready

MEASURE
performance precisely
tradeoffs honestly

DOCUMENT
learnings clearly
mistakes openly


Let's Connect

LinkedIn Medium Kaggle Twitter

tubaasid@gmail.comportfolio


sipping chai • building systems • learning deeply

views

Pinned Loading

  1. DataPath DataPath Public

    DataPath of Single Cycle Core on Scala(Chisel) based on RISC-V ISA

    Scala

  2. Mastermind Mastermind Public

    It was a mind game written in C++. Program used to create a random sequence of 8 colours and you had to guess the right sequence in 3 tries.

    C++ 1

  3. Voice-Recognition Voice-Recognition Public

    Voice recognition is the process of identifying the speaker of the audio provided in the input by matching it with the audios stored in the database. This can be used in security purposes, which me…

    MATLAB 4 3

  4. PiplinedDataPath PiplinedDataPath Public

    Piplined DataPath

    Scala

  5. Mazar-e-Quaid Mazar-e-Quaid Public

    It is a structure of Tomb of Quaid aka Mazar-e-Quaid, drawn on C++ using graphics.h library in computer graphics.

    C++ 1

  6. Smart-Traffic-Management-System Smart-Traffic-Management-System Public

    There has been a population increase which has consequently led to traffic congestion in the city of Karachi. Making a smart traffic management system that makes use of video and picture data of th…

    Python 9