From fe1ee5430f9f0e912c000a0988e99a8241d02906 Mon Sep 17 00:00:00 2001 From: "github-actions[bot]" <41898282+github-actions[bot]@users.noreply.github.com> Date: Tue, 4 Feb 2025 23:32:58 -0600 Subject: [PATCH] [ci] release (#232) Co-authored-by: github-actions[bot] --- .changeset/bright-chefs-double.md | 13 ------------- .changeset/gentle-jokes-fail.md | 13 ------------- packages/prompts/CHANGELOG.md | 27 +++++++++++++++++++++++++++ packages/prompts/package.json | 7 +++++-- 4 files changed, 32 insertions(+), 28 deletions(-) delete mode 100644 .changeset/bright-chefs-double.md delete mode 100644 .changeset/gentle-jokes-fail.md diff --git a/.changeset/bright-chefs-double.md b/.changeset/bright-chefs-double.md deleted file mode 100644 index 55d98adf..00000000 --- a/.changeset/bright-chefs-double.md +++ /dev/null @@ -1,13 +0,0 @@ ---- -'@clack/prompts': minor ---- - -Adds a new `indicator` option to `spinner`, which supports the original `"dots"` loading animation or a new `"timer"` loading animation. - -```ts -import * as p from '@clack/prompts'; - -const spin = p.spinner({ indicator: 'timer' }); -spin.start('Loading'); -await sleep(3000); -spin.stop('Loaded'); diff --git a/.changeset/gentle-jokes-fail.md b/.changeset/gentle-jokes-fail.md deleted file mode 100644 index 9ac14013..00000000 --- a/.changeset/gentle-jokes-fail.md +++ /dev/null @@ -1,13 +0,0 @@ ---- -"@clack/prompts": minor ---- - -Adds `stream` API which provides the same methods as `log`, but for iterable (even async) message streams. This is particularly useful for AI responses which are dynamically generated by LLMs. - -```ts -import * as p from '@clack/prompts'; - -await p.stream.step((async function* () { - yield* generateLLMResponse(question); -})()) -``` diff --git a/packages/prompts/CHANGELOG.md b/packages/prompts/CHANGELOG.md index 0568975d..57e0116e 100644 --- a/packages/prompts/CHANGELOG.md +++ b/packages/prompts/CHANGELOG.md @@ -1,5 +1,32 @@ # @clack/prompts +## 0.10.0 + +### Minor Changes + +- 613179d: Adds a new `indicator` option to `spinner`, which supports the original `"dots"` loading animation or a new `"timer"` loading animation. + + ```ts + import * as p from "@clack/prompts"; + + const spin = p.spinner({ indicator: "timer" }); + spin.start("Loading"); + await sleep(3000); + spin.stop("Loaded"); + ``` + +- a38b2bc: Adds `stream` API which provides the same methods as `log`, but for iterable (even async) message streams. This is particularly useful for AI responses which are dynamically generated by LLMs. + + ```ts + import * as p from "@clack/prompts"; + + await p.stream.step( + (async function* () { + yield* generateLLMResponse(question); + })() + ); + ``` + ## 0.9.1 ### Patch Changes diff --git a/packages/prompts/package.json b/packages/prompts/package.json index 18ebe4d8..59942582 100644 --- a/packages/prompts/package.json +++ b/packages/prompts/package.json @@ -1,6 +1,6 @@ { "name": "@clack/prompts", - "version": "0.9.1", + "version": "0.10.0", "type": "module", "main": "./dist/index.cjs", "module": "./dist/index.mjs", @@ -22,7 +22,10 @@ "url": "https://github.com/natemoo-re/clack/issues" }, "homepage": "https://github.com/natemoo-re/clack/tree/main/packages/prompts#readme", - "files": ["dist", "CHANGELOG.md"], + "files": [ + "dist", + "CHANGELOG.md" + ], "author": { "name": "Nate Moore", "email": "nate@natemoo.re",