Kuriko IWAI - Architect of Kernel Labs

Welcome to  Kernel Labs  by Kuriko IWAI.

A comprehensive Machine Learning frameworks and MLOps.

This website hosts a complehensive framework on the entire machine learning lifecycle - from algorithmic deep-dives to robust MLOps exercise. Explore:

  • AI Engineering Masterclass: Build eight AI systems to master LLM techniques.
  • Research & Blogs
    • Theory & FoundationLoss landscapes, optimization convergence, statistical models.
    • MLOpsEnterprise-grade engineering of ML lineage.
    • Learning ScenarioTechnical breakdown of specialized learning schemas.
    • LLM EngineeringTransformer, tokenization strategies, fine-tuning, and inference optimization.
    • Agentic AIVector DB embedding strategies, RAG, and Agentic decision logic
  • Labs: Experimentations on ML systems with walk-through tutorials and code snippets.
  • Solution: ML system and data pipeline engineering, AI audit services.

Critical Learning Paths

What's New

Understanding Vector Databases and Embedding Pipelines

Explore the mechanics of vector databases, text embedding (Dense, Sparse, Hybrid), and similarity metrics like Cosine Similarity with coding examples.

Machine LearningDeep LearningData SciencePythonAgentic AILLM

Traditional databases excel at keywords but fail at context.

To bridge the gap between structured storage and neural processing, engineers utilize Vector Databases and Vectorization.

This technical deep-dive explains how unstructured data is transformed into high-dimensional coordinates, explores the mathematical foundations of similarity scoring, and provides practical Python implementations for dense, sparse, and hybrid embedding tactics.

Continue Reading

How to Design a Production-Ready RAG System (Architecture + Tradeoffs) (2026 Edition)

Master industry-standard RAG architectures and how to architect an optimal RAG pipeline, balancing cost, latency, and precision.

Machine LearningDeep LearningAgentic AI

Vector search alone is no longer enough for enterprise AI.

While a simple NaiveRAG works for basic FAQs, complex reasoning and multi-document synthesis require specialized pipelines.

This guide dissects the six primary RAG architectures—including GraphRAG and Agentic RAG—and provides a rigorous decision framework to help you choose the right stack for your data’s complexity, reliability requirements, and budget.

Continue Reading

Engineer High-Fidelity SLM for Edge AI with Multi-stage Tuning Pipeline

Learn how to engineer high-fidelity Small Language Model (Llama 3.2 3B) with SFT, RKD, and DPO for edge deployment.

Machine LearningDeep LearningData SciencePythonLLM

Small models trade off intelligence for efficiency.

This technical deep-dive demonstrates how to bridge that gap.

By utilizing a three-phase training pipeline—Supervised Fine-Tuning (SFT), Response Knowledge Distillation (RKD), and Direct Preference Optimization (DPO)—we embed complex human traits into a small model, then deploy it across a high-throughput AWS SageMaker environment and privacy-first edge devices.

Continue Reading

Get 5-Point AI Security Checklist:

AI Engineering Masterclass

Module 3

Digital Clone: Persona Fine-Tuning & Edge Distillation

Engineered a high-fidelity interactive persona by distilling linguistic patterns from frontier models into a localized 3B parameter footprint.

unslothtrltransformersggufvllmsagemakerboto3openai

You'll Build: Edge-Native Digital Clone (Smartphone/Web)

Digital Clone: Persona Fine-Tuning & Edge Distillation

Production Goals:

  • Compress GPT 5.4 mini intelligence for edge AI.

What You'll Master:

  • Distill latent reasoning and Chain-of-Thought (CoT) capabilities from GPT-5.4 into a 3B model.
  • Engineer multi-stage tuning pipeline - SFT for grounding, RKD for logic, and DPO for stylistic parity.
  • Standardize input/output schemas using chat templates.
  • Implement 4-bit quantization (GGUF) to balance VRAM efficiency and perplexity for edge hardware.
  • Deploy via AWS SageMaker LMI/vLLM engine for paged-attention concurrency and real-time streaming.

Agentic AI framework

MIT licenseMIT licenseMIT licensePyPIPython

versionhq is a Python framework for autonomous agent networks that handle complex task automation without human interaction.

version UI dark mode
pypi package
agent network and task graph

Key Features

versionhq is a Python framework designed for automating complex, multi-step tasks using autonomous agent networks.

Users can either configure their agents and network manually or allow the system to automatically manage the process based on provided task goals.

Agent Network

When multiple agents handle a task, agents will adapt to specific network formation based on the task and network complexity.

You can specify a desired formation or allow the leader to determine it autonomously (default).

Solo AgentSupervisingSquadRandom
Formationsolosupervisorsquadrandom
Usage
  • A single agent with tools, knowledge, and memory.
  • When self-learning mode is on - it will turn into Random formation.
  • Leader agent gives directions, while sharing its knowledge and memory.
  • Subordinates can be solo agents or networks.
  • Share tasks, knowledge, and memory among network members.
  • A single agent handles tasks, asking help from other agents without sharing its memory or knowledge.
Use caseAn email agent drafts promo message for the given audience.The leader agent strategizes an outbound campaign plan and assigns components such as media mix or message creation to subordinate agents.An email agent and social media agent share the product knowledge and deploy multi-channel outbound campaign.1. An email agent drafts promo message for the given audience, asking insights on tones from other email agents which oversee other clusters. 2. An agent calls the external agent to deploy the campaign.

Kuriko IWAI

Kernel Labs Pte. Ltd.

Kuriko IWAI

Looking for Solutions?

Related Books

These books cover the wide range of ML theories and practices from fundamentals to PhD level.

Linear Algebra Done Right

Linear Algebra Done Right

Foundations of Machine Learning, second edition (Adaptive Computation and Machine Learning series)

Foundations of Machine Learning, second edition (Adaptive Computation and Machine Learning series)

Designing Data-Intensive Applications: The Big Ideas Behind Reliable, Scalable, and Maintainable Systems

Designing Data-Intensive Applications: The Big Ideas Behind Reliable, Scalable, and Maintainable Systems

Designing Machine Learning Systems: An Iterative Process for Production-Ready Applications

Designing Machine Learning Systems: An Iterative Process for Production-Ready Applications

Machine Learning Design Patterns: Solutions to Common Challenges in Data Preparation, Model Building, and MLOps

Machine Learning Design Patterns: Solutions to Common Challenges in Data Preparation, Model Building, and MLOps