mirror of
https://github.com/anomalyco/opencode.git
synced 2026-05-01 18:26:38 +00:00
refactor(skill): effectify discovery loading
This commit is contained in:
@@ -95,6 +95,7 @@
|
|||||||
"@openrouter/ai-sdk-provider": "1.5.4",
|
"@openrouter/ai-sdk-provider": "1.5.4",
|
||||||
"@opentui/core": "0.1.87",
|
"@opentui/core": "0.1.87",
|
||||||
"@opentui/solid": "0.1.87",
|
"@opentui/solid": "0.1.87",
|
||||||
|
"@effect/platform-node": "4.0.0-beta.31",
|
||||||
"@parcel/watcher": "2.5.1",
|
"@parcel/watcher": "2.5.1",
|
||||||
"@pierre/diffs": "catalog:",
|
"@pierre/diffs": "catalog:",
|
||||||
"@solid-primitives/event-bus": "1.1.2",
|
"@solid-primitives/event-bus": "1.1.2",
|
||||||
|
|||||||
@@ -1,14 +1,12 @@
|
|||||||
import path from "path"
|
import { NodeFileSystem, NodePath } from "@effect/platform-node"
|
||||||
import { Effect, Layer, Schema, ServiceMap } from "effect"
|
import { Effect, FileSystem, Layer, Path, Schema, ServiceMap } from "effect"
|
||||||
import { FetchHttpClient, HttpClient, HttpClientRequest, HttpClientResponse } from "effect/unstable/http"
|
import { FetchHttpClient, HttpClient, HttpClientRequest, HttpClientResponse } from "effect/unstable/http"
|
||||||
import { Global } from "../global"
|
import { Global } from "../global"
|
||||||
import { Log } from "../util/log"
|
import { Log } from "../util/log"
|
||||||
import { Filesystem } from "../util/filesystem"
|
|
||||||
import { withTransientReadRetry } from "@/util/effect-http-client"
|
import { withTransientReadRetry } from "@/util/effect-http-client"
|
||||||
|
|
||||||
class IndexSkill extends Schema.Class<IndexSkill>("IndexSkill")({
|
class IndexSkill extends Schema.Class<IndexSkill>("IndexSkill")({
|
||||||
name: Schema.String,
|
name: Schema.String,
|
||||||
description: Schema.String,
|
|
||||||
files: Schema.Array(Schema.String),
|
files: Schema.Array(Schema.String),
|
||||||
}) {}
|
}) {}
|
||||||
|
|
||||||
@@ -16,11 +14,8 @@ class Index extends Schema.Class<Index>("Index")({
|
|||||||
skills: Schema.Array(IndexSkill),
|
skills: Schema.Array(IndexSkill),
|
||||||
}) {}
|
}) {}
|
||||||
|
|
||||||
export namespace Discovery {
|
const skillConcurrency = 4
|
||||||
export function dir() {
|
const fileConcurrency = 8
|
||||||
return path.join(Global.Path.cache, "skills")
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
export namespace DiscoveryService {
|
export namespace DiscoveryService {
|
||||||
export interface Service {
|
export interface Service {
|
||||||
@@ -35,113 +30,89 @@ export class DiscoveryService extends ServiceMap.Service<DiscoveryService, Disco
|
|||||||
DiscoveryService,
|
DiscoveryService,
|
||||||
Effect.gen(function* () {
|
Effect.gen(function* () {
|
||||||
const log = Log.create({ service: "skill-discovery" })
|
const log = Log.create({ service: "skill-discovery" })
|
||||||
const http = withTransientReadRetry(yield* HttpClient.HttpClient)
|
const fs = yield* FileSystem.FileSystem
|
||||||
|
const path = yield* Path.Path
|
||||||
|
const http = HttpClient.filterStatusOk(withTransientReadRetry(yield* HttpClient.HttpClient))
|
||||||
|
const cache = path.join(Global.Path.cache, "skills")
|
||||||
|
|
||||||
const get = Effect.fn("DiscoveryService.get")((url: string, dest: string) =>
|
const download = Effect.fn("DiscoveryService.download")(function* (url: string, dest: string) {
|
||||||
Effect.gen(function* () {
|
if (yield* fs.exists(dest).pipe(Effect.orDie)) return true
|
||||||
if (yield* Effect.promise(() => Filesystem.exists(dest))) return true
|
|
||||||
|
|
||||||
const req = HttpClientRequest.get(url)
|
return yield* HttpClientRequest.get(url).pipe(
|
||||||
const response = yield* http.execute(req).pipe(
|
http.execute,
|
||||||
Effect.catch((err) => {
|
Effect.flatMap((res) => res.arrayBuffer),
|
||||||
|
Effect.flatMap((body) =>
|
||||||
|
fs
|
||||||
|
.makeDirectory(path.dirname(dest), { recursive: true })
|
||||||
|
.pipe(Effect.flatMap(() => fs.writeFile(dest, new Uint8Array(body)))),
|
||||||
|
),
|
||||||
|
Effect.as(true),
|
||||||
|
Effect.catch((err) =>
|
||||||
|
Effect.sync(() => {
|
||||||
log.error("failed to download", { url, err })
|
log.error("failed to download", { url, err })
|
||||||
return Effect.succeed(null)
|
return false
|
||||||
}),
|
}),
|
||||||
)
|
),
|
||||||
if (!response) return false
|
)
|
||||||
|
})
|
||||||
const ok = yield* HttpClientResponse.filterStatusOk(response).pipe(
|
|
||||||
Effect.catch(() => {
|
|
||||||
log.error("failed to download", { url, status: response.status })
|
|
||||||
return Effect.succeed(null)
|
|
||||||
}),
|
|
||||||
)
|
|
||||||
if (!ok) return false
|
|
||||||
|
|
||||||
const body = yield* ok.arrayBuffer.pipe(
|
|
||||||
Effect.catch((err) => {
|
|
||||||
log.error("failed to read download body", { url, err })
|
|
||||||
return Effect.succeed(null)
|
|
||||||
}),
|
|
||||||
)
|
|
||||||
if (!body) return false
|
|
||||||
|
|
||||||
yield* Effect.promise(() => Filesystem.write(dest, Buffer.from(body)))
|
|
||||||
return true
|
|
||||||
}),
|
|
||||||
)
|
|
||||||
|
|
||||||
const pull: DiscoveryService.Service["pull"] = Effect.fn("DiscoveryService.pull")(function* (url: string) {
|
const pull: DiscoveryService.Service["pull"] = Effect.fn("DiscoveryService.pull")(function* (url: string) {
|
||||||
const base = url.endsWith("/") ? url : `${url}/`
|
const base = url.endsWith("/") ? url : `${url}/`
|
||||||
const index = new URL("index.json", base).href
|
const index = new URL("index.json", base).href
|
||||||
const cache = Discovery.dir()
|
|
||||||
const host = base.slice(0, -1)
|
const host = base.slice(0, -1)
|
||||||
|
|
||||||
log.info("fetching index", { url: index })
|
log.info("fetching index", { url: index })
|
||||||
|
|
||||||
const req = HttpClientRequest.get(index).pipe(HttpClientRequest.acceptJson)
|
const data = yield* HttpClientRequest.get(index).pipe(
|
||||||
const response = yield* http.execute(req).pipe(
|
HttpClientRequest.acceptJson,
|
||||||
Effect.catch((err) => {
|
http.execute,
|
||||||
log.error("failed to fetch index", { url: index, err })
|
Effect.flatMap(HttpClientResponse.schemaBodyJson(Index)),
|
||||||
return Effect.succeed(null)
|
Effect.catch((err) =>
|
||||||
}),
|
Effect.sync(() => {
|
||||||
|
log.error("failed to fetch index", { url: index, err })
|
||||||
|
return null
|
||||||
|
}),
|
||||||
|
),
|
||||||
)
|
)
|
||||||
if (!response) return Array<string>()
|
|
||||||
|
|
||||||
const ok = yield* HttpClientResponse.filterStatusOk(response).pipe(
|
if (!data) return []
|
||||||
Effect.catch(() => {
|
|
||||||
log.error("failed to fetch index", { url: index, status: response.status })
|
|
||||||
return Effect.succeed(null)
|
|
||||||
}),
|
|
||||||
)
|
|
||||||
if (!ok) return Array<string>()
|
|
||||||
|
|
||||||
const data = yield* HttpClientResponse.schemaBodyJson(Index)(ok).pipe(
|
|
||||||
Effect.catch((err) => {
|
|
||||||
log.error("failed to parse index", { url: index, err })
|
|
||||||
return Effect.succeed(null)
|
|
||||||
}),
|
|
||||||
)
|
|
||||||
if (!data) {
|
|
||||||
log.warn("invalid index format", { url: index })
|
|
||||||
return Array<string>()
|
|
||||||
}
|
|
||||||
|
|
||||||
const list = data.skills.filter((skill) => {
|
const list = data.skills.filter((skill) => {
|
||||||
if (!skill.name || !Array.isArray(skill.files)) {
|
if (!skill.files.includes("SKILL.md")) {
|
||||||
log.warn("invalid skill entry", { url: index, skill })
|
log.warn("skill entry missing SKILL.md", { url: index, skill: skill.name })
|
||||||
return false
|
return false
|
||||||
}
|
}
|
||||||
return true
|
return true
|
||||||
})
|
})
|
||||||
|
|
||||||
const dirs = yield* Effect.all(
|
const dirs = yield* Effect.forEach(
|
||||||
list.map((skill) =>
|
list,
|
||||||
|
(skill) =>
|
||||||
Effect.gen(function* () {
|
Effect.gen(function* () {
|
||||||
const root = path.join(cache, skill.name)
|
const root = path.join(cache, skill.name)
|
||||||
|
|
||||||
yield* Effect.all(
|
yield* Effect.forEach(
|
||||||
skill.files.map((file) => {
|
skill.files,
|
||||||
const link = new URL(file, `${host}/${skill.name}/`).href
|
(file) => download(new URL(file, `${host}/${skill.name}/`).href, path.join(root, file)),
|
||||||
const dest = path.join(root, file)
|
{ concurrency: fileConcurrency },
|
||||||
return get(link, dest)
|
|
||||||
}),
|
|
||||||
{ concurrency: "unbounded" },
|
|
||||||
)
|
)
|
||||||
|
|
||||||
const md = path.join(root, "SKILL.md")
|
const md = path.join(root, "SKILL.md")
|
||||||
return (yield* Effect.promise(() => Filesystem.exists(md))) ? root : null
|
return (yield* fs.exists(md).pipe(Effect.orDie)) ? root : null
|
||||||
}),
|
}),
|
||||||
),
|
{ concurrency: skillConcurrency },
|
||||||
{ concurrency: "unbounded" },
|
|
||||||
)
|
)
|
||||||
|
|
||||||
return dirs.filter((dir): dir is string => Boolean(dir))
|
return dirs.filter((dir): dir is string => dir !== null)
|
||||||
})
|
})
|
||||||
|
|
||||||
return DiscoveryService.of({ pull })
|
return DiscoveryService.of({ pull })
|
||||||
}),
|
}),
|
||||||
)
|
)
|
||||||
|
|
||||||
static readonly defaultLayer = DiscoveryService.layer.pipe(Layer.provide(FetchHttpClient.layer))
|
static readonly defaultLayer = DiscoveryService.layer.pipe(
|
||||||
|
Layer.provide(FetchHttpClient.layer),
|
||||||
|
Layer.provide(NodeFileSystem.layer),
|
||||||
|
Layer.provide(NodePath.layer),
|
||||||
|
)
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -1,6 +1,7 @@
|
|||||||
import { describe, test, expect, beforeAll, afterAll } from "bun:test"
|
import { describe, test, expect, beforeAll, afterAll } from "bun:test"
|
||||||
import { Effect } from "effect"
|
import { Effect } from "effect"
|
||||||
import { Discovery, DiscoveryService } from "../../src/skill/discovery"
|
import { DiscoveryService } from "../../src/skill/discovery"
|
||||||
|
import { Global } from "../../src/global"
|
||||||
import { Filesystem } from "../../src/util/filesystem"
|
import { Filesystem } from "../../src/util/filesystem"
|
||||||
import { rm } from "fs/promises"
|
import { rm } from "fs/promises"
|
||||||
import path from "path"
|
import path from "path"
|
||||||
@@ -10,9 +11,10 @@ let server: ReturnType<typeof Bun.serve>
|
|||||||
let downloadCount = 0
|
let downloadCount = 0
|
||||||
|
|
||||||
const fixturePath = path.join(import.meta.dir, "../fixture/skills")
|
const fixturePath = path.join(import.meta.dir, "../fixture/skills")
|
||||||
|
const cacheDir = path.join(Global.Path.cache, "skills")
|
||||||
|
|
||||||
beforeAll(async () => {
|
beforeAll(async () => {
|
||||||
await rm(Discovery.dir(), { recursive: true, force: true })
|
await rm(cacheDir, { recursive: true, force: true })
|
||||||
|
|
||||||
server = Bun.serve({
|
server = Bun.serve({
|
||||||
port: 0,
|
port: 0,
|
||||||
@@ -41,7 +43,7 @@ beforeAll(async () => {
|
|||||||
|
|
||||||
afterAll(async () => {
|
afterAll(async () => {
|
||||||
server?.stop()
|
server?.stop()
|
||||||
await rm(Discovery.dir(), { recursive: true, force: true })
|
await rm(cacheDir, { recursive: true, force: true })
|
||||||
})
|
})
|
||||||
|
|
||||||
describe("Discovery.pull", () => {
|
describe("Discovery.pull", () => {
|
||||||
@@ -52,7 +54,7 @@ describe("Discovery.pull", () => {
|
|||||||
const dirs = await pull(CLOUDFLARE_SKILLS_URL)
|
const dirs = await pull(CLOUDFLARE_SKILLS_URL)
|
||||||
expect(dirs.length).toBeGreaterThan(0)
|
expect(dirs.length).toBeGreaterThan(0)
|
||||||
for (const dir of dirs) {
|
for (const dir of dirs) {
|
||||||
expect(dir).toStartWith(Discovery.dir())
|
expect(dir).toStartWith(cacheDir)
|
||||||
const md = path.join(dir, "SKILL.md")
|
const md = path.join(dir, "SKILL.md")
|
||||||
expect(await Filesystem.exists(md)).toBe(true)
|
expect(await Filesystem.exists(md)).toBe(true)
|
||||||
}
|
}
|
||||||
@@ -94,7 +96,7 @@ describe("Discovery.pull", () => {
|
|||||||
|
|
||||||
test("caches downloaded files on second pull", async () => {
|
test("caches downloaded files on second pull", async () => {
|
||||||
// clear dir and downloadCount
|
// clear dir and downloadCount
|
||||||
await rm(Discovery.dir(), { recursive: true, force: true })
|
await rm(cacheDir, { recursive: true, force: true })
|
||||||
downloadCount = 0
|
downloadCount = 0
|
||||||
|
|
||||||
// first pull to populate cache
|
// first pull to populate cache
|
||||||
|
|||||||
Reference in New Issue
Block a user