Skip to content
View back2matching's full-sized avatar

Block or report back2matching

Block user

Prevent this user from interacting with your repositories and sending you notifications. Learn more about blocking users.

You must be logged in to block users.

Maximum 250 characters. Please don't include any personal information such as legal names or email addresses. Markdown supported. This note will be visible to only you.
Report abuse

Contact GitHub support about this user’s behavior. Learn more about reporting abuse.

Report abuse
back2matching/README.md

back2matching

Typing SVG

Profile Views   Website   X   Telegram   Email


$ whoami
> Full-stack engineer building AI agents, Web3 infrastructure & real-time systems.
> Python (59%) · TypeScript (27%) · JavaScript (8%)
> 35+ repos · 23 deployed · 5,400+ contributions this year

> now

  • FlockRun — runtime for AI agent teams: scheduling, messaging, shared knowledge, real-time dashboard
  • cigoL — reverse logic engine for automated reasoning
  • matching.work — brutalist wireframe portfolio · next.js + gsap

Tech Stack


GitHub Stats    Streak

Top Languages


Activity Graph


Trophies


code. ship. repeat.

Popular repositories Loading

  1. turboquant turboquant Public

    First open-source TurboQuant KV cache compression for LLM inference. Drop-in for HuggingFace. pip install turboquant.

    Python 7 2

  2. back2matching back2matching Public

  3. kvcache-bench kvcache-bench Public

    Benchmark every KV cache compression method on your GPU. One command, real numbers. Supports Ollama + llama.cpp.

    Python

  4. quant-sim quant-sim Public

    Which quantization should I use? One command benchmarks every quant level on YOUR GPU.

    Python

  5. turboquant-vectors turboquant-vectors Public

    Compress embeddings 6x instantly with TurboQuant. First pip package using Google's TurboQuant (ICLR 2026) for vector search. 71.9% recall vs FAISS PQ 13.3%.

    Python