IntroductionIn this article, let's talk about ASP.NET Core's memory cache. ASP.NET Core Memory Cache (IMemoryCache) is a lightweight caching solution suitable for single-instance applications or local caching in distributed environments. It provides simple APIs for storing and retrieving data while supporting expiration policies, priority settings, and other features.What is Caching?From user request to database returning data, this is a long process (exaggerating a bit—usually just tens to hundreds of milliseconds). But there's more than on...
2026年4月
Technical Information PublicationI Distilled Myself into a Skill! Open Source Release
Introduction: The Distillation TrendHello everyone, I'm programmer Yupi (Fish Pi).Recently, a "distillation" trend has swept through GitHub.No, not distilling alcohol—distilling people.Colleague.skill, Ex.skill, Nuwa.skill, Boss.skill, Self.skill... Various strange distillation projects are emerging one after another. Everyone is "encapsulating" people around them into AI skill packages.Some people distilled their resigned colleagues, letting AI continue their work; some distilled their ex-partners, chatting with the AI version to reminisce;...
conhost.exe Cannot Display U+2717: Deep Dive into Windows Console Font Fallback
Introduction: The Missing Checkmark ProblemI wanted to display U+2717 (✗, BALLOT X) in Windows' traditional console (conhost). After selecting the system's default monospace font Consolas, what appeared was a question mark inside a box. I encountered this symbol while using NeoVim, which also requires Nerd Fonts—but even many fonts supporting Nerd Fonts don't support this character.The cause of the problem is obvious: Consolas doesn't contain this glyph. Simply switch to a font that includes it. After checking fileformat.info's glyph support...
When AI Agents Extend Call Chains: Latency Becomes a Business
Introduction: The Hidden Cost of LatencyMany teams only truly realize how expensive latency is after their product goes live.A seemingly simple AI Agent request often isn't just a single model call behind the scenes—it's an entire execution chain: the model understands the task, calls tools, reads data, reasons again, calls APIs, and finally generates results. Users only see one answer, but the system may have traveled back and forth between different services a dozen times.If each step adds a little waiting time, what accumulates in the end...
Deep Dive into vLLM Weight Loading: From Challenges to Ideal Architecture
Introduction: What Problems Does Weight Loading Solve?Before diving into vLLM's weight loading implementation, it's essential to understand the core challenges it addresses.Large language model weights are typically stored on disk as checkpoint files. The weight loading task seems straightforward: read these files, match tensors by name, and copy data into the model's parameters. However, three critical complexities make this far from simple.Challenge 1: Tensor Sharding and Memory Control in Tensor ParallelismvLLM supports splitting a model ...
Four Open-Source Kafka Management Tools Every Developer Should Know
IntroductionIn modern distributed architecture, Kafka serves as a core message streaming platform where stability and observability are absolutely critical. However, native command-line operations are often cumbersome and lack intuitive interfaces, presenting significant challenges for daily operations and development work. Today, we're exploring four open-source and practical Kafka management tools that can dramatically simplify your workflow and boost productivity.These tools have been battle-tested in production environments and offer uni...