Logo loFT LLC

  • person Profile
  • article Articles
      • Dagster + NATS Event-Driven Pipeline Design and Implementation
      • Rust + NATS + Dagster AI Factory: OpenAI Proxy, Idempotent Design, SSE Streaming, and Go Migration Record
      • Django 5 Travel Booking Site Generation Test with Qwen3.5-122B-A10B Local Inference
      • Why EPYC 9175F's 512MB L3 Cache Accelerates MoE Inference: Hypothesis Validation with a 1T Model
      • Why Quantization Choice Changes Everything for Hermes-4.3-36B: BF16/FP8/nvfp4 Measured Comparison
      • MiniMax-2.5 229B MoE with IQ5K Quantization on Blackwell GPU: 35 tok/s Generation, 65K Context Validation
      • The Reality of 40B Dense Models: What Running IQuest-Coder-V1-40B on CPU/GPU/Aider Actually Showed
      • MiniMax-2.5 (229B MoE) Expert Offload and Web Generation: IQ5_K to IQ3_S
      • Qwen3.5-397B IQ4_NL Measured: 22.5tok/s Average from 28 Runs, Hybrid Offload Config and 400B-Class MoE Daily Viability
      • Llama-4-Scout-17B-16E Measured: CPU Q6_K 17tok/s vs GPU nvfp4 60tok/s, Cache Strategy and 100K Context Boundary
      • 1T MoE Kimi-K2.5 CPU Inference: Thread Optimization Through Long Context Operations
      • Llama-4-Maverick-17B-128E CPU Inference: Q4_K_M vs Q8_0 Speed-Quality Trade-off Measured
      • Qwen3-Coder-Next 80B in Three Modes: BF16 CPU / IQ4_NL Hybrid / nvfp4 GPU Measured
      • GLM-4.7-Flash IQ5_K Benchmark: CPU vs Hybrid vs Full GPU Performance Comparison
      • Why DeepSeek-V3.2 Appears Slower Than Kimi-K2.5: Prompt Cache Mismatches and TG Bottleneck Analysis
      • code-tree Specification, Design Intent, and Expected Effects — LLM Context Optimization Tool
      • shelpa-mcp: Design Record of a Scrapped Virtual Pipeline
      • shelpa: Design and Lessons from a Scrapped Sandbox MCP
      • Verifying ctree Refactoring Effectiveness — Project Structure Optimization
      • Building code-tree HTML Template and Markdown Scanner — Extending to Document Formats
      • Automatic Path Error Recovery MCP Tool for Local LLMs: Building pathfinder
      • Optimizing pathfinder: Model Selection, Precision Tuning, and History Correlation Validation
      • Qwen3.5-397B Autonomous Code Generation: From Dental Clinic Sites to Django CMS Foundations
      • Bilingual AI Proofreading and Translation Prompt Definitions
      • LTX-2 Video Generation Prompt Engineering: From 36-Scene Horror to Cinematic Continuity Pipelines
    Logo
    email Contact Us
      • Japanese
    • to navigate
    • to select
    • to close
      • Home
      • Tech Memo
      On this page
      code

      Tech Memo

      Technical documentation covering infrastructure, LLM research, software tools, and development frameworks used at loFT LLC.

      info
      These articles use AI-generated summaries of Obsidian notes originally kept as technical memos.
      info
      English translations are produced with AI assistance.
      dns folder

      Infrastructure

      Server hardware, network topology, container orchestration, and monitoring stack documentation.

      dns folder

      system architecture

      Server hardware, network topology, container orchestration, and monitoring stack documentation.

      psychology folder

      LLM Research

      Large language model benchmarks, CPU/GPU inference validation, and optimization research.

      build folder

      Software Tools

      Development tools, IDE configurations, MCP integrations, and code analysis utilities.

      account_tree folder

      Workflows

      Development workflows, coding philosophy, AI agent configurations, and automation practices.


      © 2017-2026 loFT LLC