mirror of
https://github.com/affaan-m/everything-claude-code.git
synced 2026-04-05 00:33:27 +08:00
* docs(zh-CN): sync Chinese docs with latest upstream changes * update --------- Co-authored-by: neo <neo.dowithless@gmail.com>
162 lines
5.5 KiB
Markdown
162 lines
5.5 KiB
Markdown
---
|
||
name: content-hash-cache-pattern
|
||
description: 使用SHA-256内容哈希缓存昂贵的文件处理结果——路径无关、自动失效、服务层分离。
|
||
origin: ECC
|
||
---
|
||
|
||
# 内容哈希文件缓存模式
|
||
|
||
使用 SHA-256 内容哈希作为缓存键,缓存昂贵的文件处理结果(PDF 解析、文本提取、图像分析)。与基于路径的缓存不同,此方法在文件移动/重命名后仍然有效,并在内容更改时自动失效。
|
||
|
||
## 何时激活
|
||
|
||
* 构建文件处理管道时(PDF、图像、文本提取)
|
||
* 处理成本高且同一文件被重复处理时
|
||
* 需要一个 `--cache/--no-cache` CLI 选项时
|
||
* 希望在不修改现有纯函数的情况下为其添加缓存时
|
||
|
||
## 核心模式
|
||
|
||
### 1. 基于内容哈希的缓存键
|
||
|
||
使用文件内容(而非路径)作为缓存键:
|
||
|
||
```python
|
||
import hashlib
|
||
from pathlib import Path
|
||
|
||
_HASH_CHUNK_SIZE = 65536 # 64KB chunks for large files
|
||
|
||
def compute_file_hash(path: Path) -> str:
|
||
"""SHA-256 of file contents (chunked for large files)."""
|
||
if not path.is_file():
|
||
raise FileNotFoundError(f"File not found: {path}")
|
||
sha256 = hashlib.sha256()
|
||
with open(path, "rb") as f:
|
||
while True:
|
||
chunk = f.read(_HASH_CHUNK_SIZE)
|
||
if not chunk:
|
||
break
|
||
sha256.update(chunk)
|
||
return sha256.hexdigest()
|
||
```
|
||
|
||
**为什么使用内容哈希?** 文件重命名/移动 = 缓存命中。内容更改 = 自动失效。无需索引文件。
|
||
|
||
### 2. 用于缓存条目的冻结数据类
|
||
|
||
```python
|
||
from dataclasses import dataclass
|
||
|
||
@dataclass(frozen=True, slots=True)
|
||
class CacheEntry:
|
||
file_hash: str
|
||
source_path: str
|
||
document: ExtractedDocument # The cached result
|
||
```
|
||
|
||
### 3. 基于文件的缓存存储
|
||
|
||
每个缓存条目都存储为 `{hash}.json` —— 通过哈希实现 O(1) 查找,无需索引文件。
|
||
|
||
```python
|
||
import json
|
||
from typing import Any
|
||
|
||
def write_cache(cache_dir: Path, entry: CacheEntry) -> None:
|
||
cache_dir.mkdir(parents=True, exist_ok=True)
|
||
cache_file = cache_dir / f"{entry.file_hash}.json"
|
||
data = serialize_entry(entry)
|
||
cache_file.write_text(json.dumps(data, ensure_ascii=False), encoding="utf-8")
|
||
|
||
def read_cache(cache_dir: Path, file_hash: str) -> CacheEntry | None:
|
||
cache_file = cache_dir / f"{file_hash}.json"
|
||
if not cache_file.is_file():
|
||
return None
|
||
try:
|
||
raw = cache_file.read_text(encoding="utf-8")
|
||
data = json.loads(raw)
|
||
return deserialize_entry(data)
|
||
except (json.JSONDecodeError, ValueError, KeyError):
|
||
return None # Treat corruption as cache miss
|
||
```
|
||
|
||
### 4. 服务层包装器(单一职责原则)
|
||
|
||
保持处理函数的纯净性。将缓存作为一个单独的服务层添加。
|
||
|
||
```python
|
||
def extract_with_cache(
|
||
file_path: Path,
|
||
*,
|
||
cache_enabled: bool = True,
|
||
cache_dir: Path = Path(".cache"),
|
||
) -> ExtractedDocument:
|
||
"""Service layer: cache check -> extraction -> cache write."""
|
||
if not cache_enabled:
|
||
return extract_text(file_path) # Pure function, no cache knowledge
|
||
|
||
file_hash = compute_file_hash(file_path)
|
||
|
||
# Check cache
|
||
cached = read_cache(cache_dir, file_hash)
|
||
if cached is not None:
|
||
logger.info("Cache hit: %s (hash=%s)", file_path.name, file_hash[:12])
|
||
return cached.document
|
||
|
||
# Cache miss -> extract -> store
|
||
logger.info("Cache miss: %s (hash=%s)", file_path.name, file_hash[:12])
|
||
doc = extract_text(file_path)
|
||
entry = CacheEntry(file_hash=file_hash, source_path=str(file_path), document=doc)
|
||
write_cache(cache_dir, entry)
|
||
return doc
|
||
```
|
||
|
||
## 关键设计决策
|
||
|
||
| 决策 | 理由 |
|
||
|----------|-----------|
|
||
| SHA-256 内容哈希 | 与路径无关,内容更改时自动失效 |
|
||
| `{hash}.json` 文件命名 | O(1) 查找,无需索引文件 |
|
||
| 服务层包装器 | 单一职责原则:提取功能保持纯净,缓存是独立的关注点 |
|
||
| 手动 JSON 序列化 | 完全控制冻结数据类的序列化 |
|
||
| 损坏时返回 `None` | 优雅降级,在下次运行时重新处理 |
|
||
| `cache_dir.mkdir(parents=True)` | 在首次写入时惰性创建目录 |
|
||
|
||
## 最佳实践
|
||
|
||
* **哈希内容,而非路径** —— 路径会变,内容标识不变
|
||
* 对大文件进行哈希时**分块处理** —— 避免将整个文件加载到内存中
|
||
* **保持处理函数的纯净性** —— 它们不应了解任何关于缓存的信息
|
||
* **记录缓存命中/未命中**,并使用截断的哈希值以便调试
|
||
* **优雅地处理损坏** —— 将无效的缓存条目视为未命中,永不崩溃
|
||
|
||
## 应避免的反模式
|
||
|
||
```python
|
||
# BAD: Path-based caching (breaks on file move/rename)
|
||
cache = {"/path/to/file.pdf": result}
|
||
|
||
# BAD: Adding cache logic inside the processing function (SRP violation)
|
||
def extract_text(path, *, cache_enabled=False, cache_dir=None):
|
||
if cache_enabled: # Now this function has two responsibilities
|
||
...
|
||
|
||
# BAD: Using dataclasses.asdict() with nested frozen dataclasses
|
||
# (can cause issues with complex nested types)
|
||
data = dataclasses.asdict(entry) # Use manual serialization instead
|
||
```
|
||
|
||
## 适用场景
|
||
|
||
* 文件处理管道(PDF 解析、OCR、文本提取、图像分析)
|
||
* 受益于 `--cache/--no-cache` 选项的 CLI 工具
|
||
* 跨多次运行出现相同文件的批处理
|
||
* 在不修改现有纯函数的情况下为其添加缓存
|
||
|
||
## 不适用场景
|
||
|
||
* 必须始终保持最新的数据(实时数据流)
|
||
* 缓存条目可能极其庞大的情况(应考虑使用流式处理)
|
||
* 结果依赖于文件内容之外参数的情况(例如,不同的提取配置)
|