* refactor: multi-channel infrastructure with explicit channel/is_group tracking - Add channels[] array and findChannel() routing in index.ts, replacing hardcoded whatsapp.* calls with channel-agnostic callbacks - Add channel TEXT and is_group INTEGER columns to chats table with COALESCE upsert to protect existing values from null overwrites - is_group defaults to 0 (safe: unknown chats excluded from groups) - WhatsApp passes explicit channel='whatsapp' and isGroup to onChatMetadata - getAvailableGroups filters on is_group instead of JID pattern matching - findChannel logs warnings instead of silently dropping unroutable JIDs - Migration backfills channel/is_group from JID patterns for existing DBs Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com> * feat: skills engine v0.1 — deterministic skill packages with rerere resolution Three-way merge engine for applying skill packages on top of a core codebase. Skills declare which files they add/modify, and the engine uses git merge-file for conflict detection with git rerere for automatic resolution of previously-seen conflicts. Key components: - apply: three-way merge with backup/rollback safety net - replay: clean-slate replay for uninstall and rebase - update: core version updates with deletion detection - rebase: bake applied skills into base (one-way) - manifest: validation with path traversal protection - resolution-cache: pre-computed rerere resolutions - structured: npm deps, env vars, docker-compose merging - CI: per-skill test matrix with conflict detection 151 unit tests covering merge, rerere, backup, replay, uninstall, update, rebase, structured ops, and edge cases. Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com> * feat: add Discord and Telegram skill packages Skill packages for adding Discord and Telegram channels to NanoClaw. Each package includes: - Channel implementation (add/src/channels/) - Three-way merge targets for index.ts, config.ts, routing.test.ts - Intent docs explaining merge invariants - Standalone integration tests - manifest.yaml with dependency/conflict declarations Applied via: npx tsx scripts/apply-skill.ts .claude/skills/add-discord These are inert until applied — no runtime impact. Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com> * remove unused docs (skills-system-status, implementation-guide) Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com> --------- Co-authored-by: Claude Opus 4.6 <noreply@anthropic.com>
294 lines
9.2 KiB
TypeScript
294 lines
9.2 KiB
TypeScript
import { execFileSync, execSync } from 'child_process';
|
|
import crypto from 'crypto';
|
|
import fs from 'fs';
|
|
import os from 'os';
|
|
import path from 'path';
|
|
|
|
import { clearBackup, createBackup, restoreBackup } from './backup.js';
|
|
import { BASE_DIR, NANOCLAW_DIR } from './constants.js';
|
|
import { copyDir } from './fs-utils.js';
|
|
import { acquireLock } from './lock.js';
|
|
import {
|
|
cleanupMergeState,
|
|
isGitRepo,
|
|
mergeFile,
|
|
runRerere,
|
|
setupRerereAdapter,
|
|
} from './merge.js';
|
|
import { clearAllResolutions } from './resolution-cache.js';
|
|
import { computeFileHash, readState, writeState } from './state.js';
|
|
import type { RebaseResult } from './types.js';
|
|
|
|
function walkDir(dir: string, root: string): string[] {
|
|
const results: string[] = [];
|
|
if (!fs.existsSync(dir)) return results;
|
|
|
|
for (const entry of fs.readdirSync(dir, { withFileTypes: true })) {
|
|
const fullPath = path.join(dir, entry.name);
|
|
if (entry.isDirectory()) {
|
|
results.push(...walkDir(fullPath, root));
|
|
} else {
|
|
results.push(path.relative(root, fullPath));
|
|
}
|
|
}
|
|
return results;
|
|
}
|
|
|
|
function collectTrackedFiles(
|
|
state: ReturnType<typeof readState>,
|
|
): Set<string> {
|
|
const tracked = new Set<string>();
|
|
|
|
for (const skill of state.applied_skills) {
|
|
for (const relPath of Object.keys(skill.file_hashes)) {
|
|
tracked.add(relPath);
|
|
}
|
|
}
|
|
|
|
if (state.custom_modifications) {
|
|
for (const mod of state.custom_modifications) {
|
|
for (const relPath of mod.files_modified) {
|
|
tracked.add(relPath);
|
|
}
|
|
}
|
|
}
|
|
|
|
return tracked;
|
|
}
|
|
|
|
export async function rebase(newBasePath?: string): Promise<RebaseResult> {
|
|
const projectRoot = process.cwd();
|
|
const state = readState();
|
|
|
|
if (state.applied_skills.length === 0) {
|
|
return {
|
|
success: false,
|
|
filesInPatch: 0,
|
|
error: 'No skills applied. Nothing to rebase.',
|
|
};
|
|
}
|
|
|
|
const releaseLock = acquireLock();
|
|
|
|
try {
|
|
const trackedFiles = collectTrackedFiles(state);
|
|
const baseAbsDir = path.join(projectRoot, BASE_DIR);
|
|
|
|
// Include base dir files
|
|
const baseFiles = walkDir(baseAbsDir, baseAbsDir);
|
|
for (const f of baseFiles) {
|
|
trackedFiles.add(f);
|
|
}
|
|
|
|
// Backup
|
|
const filesToBackup: string[] = [];
|
|
for (const relPath of trackedFiles) {
|
|
const absPath = path.join(projectRoot, relPath);
|
|
if (fs.existsSync(absPath)) filesToBackup.push(absPath);
|
|
const baseFilePath = path.join(baseAbsDir, relPath);
|
|
if (fs.existsSync(baseFilePath)) filesToBackup.push(baseFilePath);
|
|
}
|
|
const stateFilePath = path.join(projectRoot, NANOCLAW_DIR, 'state.yaml');
|
|
filesToBackup.push(stateFilePath);
|
|
createBackup(filesToBackup);
|
|
|
|
try {
|
|
// Generate unified diff: base vs working tree (archival record)
|
|
let combinedPatch = '';
|
|
let filesInPatch = 0;
|
|
|
|
for (const relPath of trackedFiles) {
|
|
const basePath = path.join(baseAbsDir, relPath);
|
|
const workingPath = path.join(projectRoot, relPath);
|
|
|
|
const oldPath = fs.existsSync(basePath) ? basePath : '/dev/null';
|
|
const newPath = fs.existsSync(workingPath) ? workingPath : '/dev/null';
|
|
|
|
if (oldPath === '/dev/null' && newPath === '/dev/null') continue;
|
|
|
|
try {
|
|
const diff = execFileSync('diff', ['-ruN', oldPath, newPath], {
|
|
encoding: 'utf-8',
|
|
});
|
|
if (diff.trim()) {
|
|
combinedPatch += diff;
|
|
filesInPatch++;
|
|
}
|
|
} catch (err: unknown) {
|
|
const execErr = err as { status?: number; stdout?: string };
|
|
if (execErr.status === 1 && execErr.stdout) {
|
|
combinedPatch += execErr.stdout;
|
|
filesInPatch++;
|
|
} else {
|
|
throw err;
|
|
}
|
|
}
|
|
}
|
|
|
|
// Save combined patch
|
|
const patchPath = path.join(
|
|
projectRoot,
|
|
NANOCLAW_DIR,
|
|
'combined.patch',
|
|
);
|
|
fs.writeFileSync(patchPath, combinedPatch, 'utf-8');
|
|
|
|
if (newBasePath) {
|
|
// --- Rebase with new base: three-way merge with resolution model ---
|
|
|
|
// Save current working tree content before overwriting
|
|
const savedContent: Record<string, string> = {};
|
|
for (const relPath of trackedFiles) {
|
|
const workingPath = path.join(projectRoot, relPath);
|
|
if (fs.existsSync(workingPath)) {
|
|
savedContent[relPath] = fs.readFileSync(workingPath, 'utf-8');
|
|
}
|
|
}
|
|
|
|
const absNewBase = path.resolve(newBasePath);
|
|
|
|
// Replace base
|
|
if (fs.existsSync(baseAbsDir)) {
|
|
fs.rmSync(baseAbsDir, { recursive: true, force: true });
|
|
}
|
|
fs.mkdirSync(baseAbsDir, { recursive: true });
|
|
copyDir(absNewBase, baseAbsDir);
|
|
|
|
// Copy new base to working tree
|
|
copyDir(absNewBase, projectRoot);
|
|
|
|
// Three-way merge per file: new-base ← old-base → saved-working-tree
|
|
const mergeConflicts: string[] = [];
|
|
|
|
for (const relPath of trackedFiles) {
|
|
const newBaseSrc = path.join(absNewBase, relPath);
|
|
const currentPath = path.join(projectRoot, relPath);
|
|
const saved = savedContent[relPath];
|
|
|
|
if (!saved) continue; // No working tree content to merge
|
|
if (!fs.existsSync(newBaseSrc)) {
|
|
// File only existed in working tree, not in new base — restore it
|
|
fs.mkdirSync(path.dirname(currentPath), { recursive: true });
|
|
fs.writeFileSync(currentPath, saved);
|
|
continue;
|
|
}
|
|
|
|
const newBaseContent = fs.readFileSync(newBaseSrc, 'utf-8');
|
|
if (newBaseContent === saved) continue; // No diff
|
|
|
|
// Find old base content from backup
|
|
const oldBasePath = path.join(
|
|
projectRoot,
|
|
'.nanoclaw',
|
|
'backup',
|
|
BASE_DIR,
|
|
relPath,
|
|
);
|
|
if (!fs.existsSync(oldBasePath)) {
|
|
// No old base — keep saved content
|
|
fs.writeFileSync(currentPath, saved);
|
|
continue;
|
|
}
|
|
|
|
// Save "ours" (new base content) before merge overwrites it
|
|
const oursContent = newBaseContent;
|
|
|
|
// Three-way merge: current(new base) ← old-base → saved(modifications)
|
|
const tmpSaved = path.join(
|
|
os.tmpdir(),
|
|
`nanoclaw-rebase-${crypto.randomUUID()}-${path.basename(relPath)}`,
|
|
);
|
|
fs.writeFileSync(tmpSaved, saved);
|
|
|
|
const result = mergeFile(currentPath, oldBasePath, tmpSaved);
|
|
fs.unlinkSync(tmpSaved);
|
|
|
|
if (!result.clean) {
|
|
// Try rerere resolution (three-level model)
|
|
if (isGitRepo()) {
|
|
const baseContent = fs.readFileSync(oldBasePath, 'utf-8');
|
|
setupRerereAdapter(relPath, baseContent, oursContent, saved);
|
|
const autoResolved = runRerere(currentPath);
|
|
|
|
if (autoResolved) {
|
|
execFileSync('git', ['add', relPath], { stdio: 'pipe' });
|
|
execSync('git rerere', { stdio: 'pipe' });
|
|
cleanupMergeState(relPath);
|
|
continue;
|
|
}
|
|
|
|
cleanupMergeState(relPath);
|
|
}
|
|
|
|
// Unresolved — conflict markers remain in working tree
|
|
mergeConflicts.push(relPath);
|
|
}
|
|
}
|
|
|
|
if (mergeConflicts.length > 0) {
|
|
// Return with backup pending for Claude Code / user resolution
|
|
return {
|
|
success: false,
|
|
patchFile: patchPath,
|
|
filesInPatch,
|
|
mergeConflicts,
|
|
backupPending: true,
|
|
error: `Merge conflicts in: ${mergeConflicts.join(', ')}. Resolve manually then call clearBackup(), or restoreBackup() + clearBackup() to abort.`,
|
|
};
|
|
}
|
|
} else {
|
|
// --- Rebase without new base: flatten into base ---
|
|
// Update base to current working tree state (all skills baked in)
|
|
for (const relPath of trackedFiles) {
|
|
const workingPath = path.join(projectRoot, relPath);
|
|
const basePath = path.join(baseAbsDir, relPath);
|
|
|
|
if (fs.existsSync(workingPath)) {
|
|
fs.mkdirSync(path.dirname(basePath), { recursive: true });
|
|
fs.copyFileSync(workingPath, basePath);
|
|
} else if (fs.existsSync(basePath)) {
|
|
// File was removed by skills — remove from base too
|
|
fs.unlinkSync(basePath);
|
|
}
|
|
}
|
|
}
|
|
|
|
// Update state
|
|
const now = new Date().toISOString();
|
|
|
|
for (const skill of state.applied_skills) {
|
|
const updatedHashes: Record<string, string> = {};
|
|
for (const relPath of Object.keys(skill.file_hashes)) {
|
|
const absPath = path.join(projectRoot, relPath);
|
|
if (fs.existsSync(absPath)) {
|
|
updatedHashes[relPath] = computeFileHash(absPath);
|
|
}
|
|
}
|
|
skill.file_hashes = updatedHashes;
|
|
}
|
|
|
|
delete state.custom_modifications;
|
|
state.rebased_at = now;
|
|
writeState(state);
|
|
|
|
// Clear stale resolution cache (base has changed, old resolutions invalid)
|
|
clearAllResolutions(projectRoot);
|
|
|
|
clearBackup();
|
|
|
|
return {
|
|
success: true,
|
|
patchFile: patchPath,
|
|
filesInPatch,
|
|
rebased_at: now,
|
|
};
|
|
} catch (err) {
|
|
restoreBackup();
|
|
clearBackup();
|
|
throw err;
|
|
}
|
|
} finally {
|
|
releaseLock();
|
|
}
|
|
}
|