Hualin Luan Cloud Native · Quant Trading · AI Engineering

Articles

Articles

Browse English seed articles with localized filters and route-safe links.

Article browser

Browse articles

Browse localized articles with card, list, or compact layouts. The view preference is stored only in this browser and locale.

Article filters

Filter articles

Showing 72 / 72 articles

Current view: Card

Results: 72 / 72

PinnedArticleMicroservice governance

From enterprise-level CF platform to cloud native (1): Architect's review - the gains and losses of microservice governance in the era of enterprise-level CF platform

Based on the front-line architecture practice of enterprise-level CF platforms from 2015 to 2020 and industry observations from 2015 to 2026 (to date), we review the microservice governance design decisions in the Cloud Foundry era and analyze which ones have withstood the test of time and which ones have been reconstructed by the cloud native wave.

MicroservicesCloud FoundryArchitectureGovernanceSpring Cloud
PinnedArticleMicroservice governance

From enterprise-level CF platform to cloud native (2): Observability-driven governance—from monitoring large screens to precise decision-making systems

With 6 years of practical experience as an enterprise-level platform architect, we analyze the core position of observability in microservice governance, from data islands to OpenTelemetry unified standards, and build a governance system for accurate decision-making.

ObservabilityOpentelemetryMicroservicesGovernanceMonitoring
PinnedArticleMicroservice governance

From enterprise-level CF platform to cloud native (3): The evolution of traffic management - from Spring Cloud Gateway to Gateway API and Ambient Mesh

Review the practice of Spring Cloud Gateway in the enterprise-level CF platform, analyze the standardization value of Kubernetes Gateway API, explore the evolution logic from Service Mesh to Ambient Mesh, and provide a decision-making framework for enterprise traffic management selection.

MicroservicesTraffic ManagementSpring Cloud GatewayGateway ApiService Mesh
PinnedArticleMicroservice governance

From enterprise-level CF platform to cloud native (4): Redefining elastic fault tolerance—from Hystrix to adaptive governance

Review Hystrix's historical position in microservice elastic governance, analyze Resilience4j's lightweight design philosophy, explore new paradigms of adaptive fault tolerance and chaos engineering, and provide practical guidance for enterprises to build resilient systems.

MicroservicesResilienceCircuit BreakerHystrixResilience4j
PinnedArticleMicroservice governance

From enterprise-level CF platform to cloud native (5): The evolution of release governance—from manual approval to progressive delivery

Review the manual approval model of traditional release governance, analyze the evolution of blue-green deployment and canary release, explore the new paradigm of GitOps and progressive delivery, and provide practical guidance for enterprises to build an efficient and secure release system.

MicroservicesRelease GovernanceBlue GreenCanaryFeature Flags
PinnedArticleMicroservice governance

From enterprise-level CF platform to cloud native (6): Summary—an architect’s perspective on enterprise-level microservice governance

Review the evolution of microservice governance over the past ten years from 2015 to 2026 (to date), refine the first principles of architects, summarize the implementation paths and common pitfalls of enterprise-level governance, look forward to future trends, and provide a systematic thinking framework for technical decision-makers.

MicroservicesGovernanceArchitectureCloud NativeEnterprise
ArticleAI programming assessment

Why do you need to be a coding mentor for AI?

When AI programming assistants become standard equipment, the real competitiveness is no longer whether they can use AI, but whether they can judge, calibrate and constrain the engineering output of AI. This article starts from trust gaps, feedback protocols, evaluation standards and closed-loop capabilities to establish the core framework of "Humans as Coding Mentors".

Ai Coding MentorProgramming EvaluationHuman Ai CollaborationOriginal Interpretation
ArticleAI programming assessment

Panorama of AI programming ability evaluation: from HumanEval to SWE-bench, the evolution and selection of benchmarks

Public benchmarks are not a decoration for model rankings, but a measurement tool for understanding the boundaries of AI programming capabilities. This article starts from benchmarks such as HumanEval, APPS, CodeContests, SWE-bench, LiveCodeBench and Aider, and explains how to read the rankings, how to choose benchmarks, and how to convert public evaluations into the team's own Coding Mentor evaluation system.

Ai Coding MentorProgramming BenchmarkOriginal InterpretationHuman EvalSwe Bench
ArticleAI programming assessment

How to design high-quality programming questions: from question surface to evaluation contract

High-quality programming questions are not longer prompts, but assessment contracts that can stably expose the boundaries of abilities. This article starts from Bloom level, difficulty calibration, task contract, test design and question bank management to explain how to build a reproducible question system for AI Coding Mentor.

Ai Coding MentorProblem DesignOriginal InterpretationCoding ExercisesBloom Taxonomy
ArticleAI programming assessment

Four-step approach to AI capability assessment: from one test to continuous system evaluation

Serving as a coding mentor for AI is not about doing a model evaluation, but establishing an evaluation operation system that can continuously expose the boundaries of capabilities, record failure evidence, drive special improvements, and support collaborative decision-making.

Ai Coding MentorEvaluation MethodologyOriginal InterpretationBaseline TestingContinuous Assessment
ArticleAI programming assessment

Practical cases: feedback protocol, evaluation closed loop, code review and programming education data

Case studies should not stop at “how to use AI tools better”. This article uses four engineering scenarios: model selection evaluation, feedback protocol design, code review signal precipitation, and programming education data closed loop to explain how humans can transform the AI ​​collaboration process into evaluable, trainable, and reusable mentor signals.

Ai Coding MentorCase StudyOriginal InterpretationFeedback ProtocolEvaluation Framework
ArticleAI programming assessment

From delivery to training: How to turn AI programming collaboration into a Coding Mentor data closed loop

The real organizational value of AI programming assistants is not just to increase delivery speed, but to precipitate trainable, evaluable, and reusable mentor signals in every requirement disassembly, code generation, review and revision, test verification, and online review. This article reconstructs the closed-loop framework of AI training, AI-assisted product engineering delivery, high-quality SFT data precipitation, and model evaluation.

Ai Coding MentorEvaluation SystemOriginal InterpretationData FlywheelAI Engineering
ArticleEval Harness

A truly mature Eval Harness will not just focus on the answer

If an eval harness can only tell you the success or failure of a task, but cannot explain whether the agent called the correct capabilities, in what environment it was executed, why it failed, and why it succeeded, then what it gives is not a systematic judgment, but just a score card. This article is based on LangChain's discussion of skills eval and extends my complete understanding of artifact-based scoring, invocation metrics, trace design, workflow eval and evaluation histology.

EvalsAgent SkillsLangsmithTracingAgents
ArticleAgent Harness

Agent Harness is not a supporting role, but the most underrated main battleground of AI engineering in 2026

What really determines the upper limit of an agent is often not the model itself, but the harness organized around the model. This article is based on LangChain's disassembly of the agent harness, extending my complete understanding of file systems, code execution, context management, verification closed loops and long-term task endurance. It also explains why the focus of AI engineering competition in 2026 is shifting from 'model capabilities' to 'working system design'.

AgentsHarnessContext EngineeringAI EngineeringLangchain
ArticleMCP Runtime

Agent Runtime does not have to be local, Colab MCP gives a more realistic direction

The value of Colab MCP is not only to run Python on the cloud, but also to turn the agent's execution environment into a notebook space that is visible, editable, and can continue to work. For many tasks, what really matters is not the remote execution itself, but how the remote artifact supports human-machine collaboration. This article is based on Google's introduction to Colab MCP Server and extends my complete understanding of runtime surface, artifact-centered design, remote workbench and visibility trust mechanism.

McpColabRuntimeNotebooksGoogle
ArticleEval Harness

The most misleading thing about Agent Benchmark is not the model score, but the infrastructure noise.

In agentic coding eval, the model is not the only variable. Resource headroom, kill semantics, concurrency pressure, network status, and sandbox behavior can all change task results. If these conditions are not transparent, small margins on the leaderboard are often less telling than they seem. This article is based on Anthropic's analysis of infrastructure noise and extends my complete understanding of agent benchmark interpretability, disclosure discipline, repeated experiments, and system-level evaluation perspectives.

EvalsInfrastructureBenchmarkAgentsAnthropic
ArticleMCP Runtime

What MCP changes is not tool access, but the cost structure of Agents.

The real significance of MCP is not just to unify tool access, but to move a large number of intermediate processes that should be handled by the runtime out of the expensive LLM cycle. What it changes is not 'how many tools can be connected', but how the agent uses context, code execution and runtime control flow. This article is based on Anthropic's discussion of code execution with MCP and extends my complete understanding of direct tool-calling, progressive disclosure, runtime economics and executable skills.

McpCode ExecutionContext EngineeringAgentsAnthropic
ArticleAgent Harness

What the long-term task agent really lacks is not intelligence, but the handover, recovery and acceptance capabilities.

The failure of long-term task agents often does not stem from the model's inability to think, but from the system's failure to design 'handover, recovery, verification, and continuation' as first-class citizens. This article is based on Anthropic's discussion of long-running agent harness, extending my complete views on cross-session execution, state externalization, feature contract, smoke test, browser verification and multi-round execution structure. It also explains why a truly usable agent does not run for a long time at a time, but can catch it round after round.

AgentsLong Running AgentsHarnessAnthropicVerification
ArticleOpenClaw security in-depth interpretation

Original interpretation: Why do OpenClaw security incidents always happen after 'the risk is already known'?

Why do OpenClaw security incidents always happen after 'the risk is already known'? This article does not blame the model for being out of control, but instead asks about the design flaws of execution rights: when the system puts execution rights, audit rights, and rollback rights on the same link, how does organizational blindness amplify controllable deviations into accidents step by step?

Original InterpretationOpenclawAgent SecurityIncident Review
ArticleOpenClaw security in-depth interpretation

Original interpretation: Why is the lightweight Agent solution likely to be closer to production reality than the 'big and comprehensive' solution?

This is not a chicken soup article praising 'lightweight', but an article against engineering illusion: many OpenClaw Agent stacks that appear to be stronger only front-load complexity into demonstration capabilities, but rearrange the cost into production failures and early morning duty costs.

Original InterpretationOpenclawNanobotContrarian
ArticleOpenClaw security in-depth interpretation

Original interpretation: Treat Notion as the control plane of 18 Agents. The first thing to solve is never 'automation'

This article does not discuss whether the console interface is good-looking or not, but discusses a more fundamental production issue: when you connect 18 OpenClaw Agents to the Notion control plane, is the system amplifying team productivity, or is it amplifying scheduling noise and status chaos?

Original InterpretationOpenclawMulti AgentOperator Playbook
ArticleOpenClaw security in-depth interpretation

Original interpretation: Putting Agent into ESP32, the easiest thing to avoid is not the performance pit, but the boundary illusion.

This article does not describe the ESP32 Edge Agent as a cool technology trial, but dismantles the four most common misunderstandings: running the board does not mean the system is usable, being offline is not just a network problem, and local success does not mean on-site maintainability. Edge deployments require new engineering assumptions.

Original InterpretationOpenclawEsp32Edge Agent
ArticleOpenClaw security in-depth interpretation

Original interpretation: When OpenClaw costs get out of control, the first thing to break is never the unit price, but the judgment framework.

If OpenClaw API fee control only focuses on the unit price of the model, it will usually turn into an illusion of cheapness in the end: the book will look good in the short term, but structural waste will still quietly accumulate in the background. This paper reconstructs a cost framework including budget boundaries, task layering and entry routing.

Original InterpretationOpenclawFinopsFramework
ArticleOpenClaw security in-depth interpretation

Original interpretation: When the Agent tries to 'take away the password', what is exposed is never just a leak point

Rewrite 'Agent knows your password' into a more uncomfortable accident review: the real failure is not a certain encryption action, but the team's use of credentials as a default capability that is always online, constantly visible, and constantly callable. This article discusses runtime governance gaps.

Original InterpretationOpenclawCredentialsIncident Review
ArticleOpenClaw security in-depth interpretation

Original interpretation: Why what OpenClaw really lacks is not more prompt words, but a tool firewall that dares to say 'no'

Many teams pin OpenClaw safety on prompt constraints, but what really determines the upper limit of accidents is not what the model thinks, but whether the system allows the model's ideas to be directly turned into tool execution. This article proposes a four-layer governance framework of 'intention-adjudication-execution-audit'.

Original InterpretationOpenclawTool FirewallFramework
ArticleOpenClaw security in-depth interpretation

Original interpretation: It is not difficult to deploy OpenClaw to AWS. The difficulty is not to mistake 'repeatable deployment' for 'already safe'

Dispel a very common but dangerous illusion: when teams say 'we've reinforced it with Terraform', they often just complete the starting point, but mistakenly believe that they are at the end. IaC can make deployment consistent, but it cannot automatically make OpenClaw systems continuously secure.

Original InterpretationOpenclawTerraformSecurity
ArticleOpenClaw security in-depth interpretation

Original interpretation: The real priority for Agent credential security is not 'where to put it', but 'who can touch it and when'

Refuting an all-too-common misconception: OpenClaw credential security is complete as long as key escrow, encrypted storage, and rotation are done. The reality is just the opposite. The most likely place for trouble often occurs at runtime - not 'where' it is placed, but 'who can touch it and when'.

Original InterpretationOpenclawClawshellContrarian
ArticleOpenClaw security in-depth interpretation

Original interpretation: Looking at the three types of OpenClaw security articles together, it is not the vulnerabilities that are really revealed, but the lag in governance.

When the three topics of prompt word injection, credential leakage, and tool firewalls are put on the same table, you will find that they point to the same core contradiction: OpenClaw's capabilities are expanding faster than execution rights management. This article synthesizes the common conclusions of three security articles.

Original InterpretationOpenclawPrompt InjectionSynthesis