About a year ago, I kicked off (Waver), a project designed to analyze codebases and build developer friendly, readable tutorials. The goal was to generate clean Markdown and Mermaid diagrams so onboarding becomes frictionless and the docs can slide right into existing project pipelines.
I chose Langchain4J for the LLM integration. And in a classic "engineers-quest", I actually ended up building JGraphlet as well, which came to life while I was trying to squeeze better performance out of the LLM communication.
Working in pure Java offers clear benefits, specifically regarding granular control and performance optimization. However, these implementations often become highly opinionated. In contrast, while SKILLS maintains a degree of opinionated design, its actual application depends entirely on the coding agent. Both methods have advantages, and I’m using this post to process what I’ve learned while writing my first SKILL. Here is a code example from my Java project.
package dev.shaaf.waver.cli;
import dev.langchain4j.model.chat.ChatModel;
import dev.shaaf.jgraphlet.TaskPipeline;
import dev.shaaf.waver.llm.config.AppConfig;
import dev.shaaf.waver.llm.config.ModelProviderFactory;
import dev.shaaf.waver.llm.tutorial.task.*;
import java.nio.file.Path;
import java.nio.file.Paths;
import java.util.logging.Logger;
public class TutorialGenerator {
private static final Logger logger = Logger.getLogger(TutorialGenerator.class.getName());
public static void generate(AppConfig appConfig) {
ChatModel chatModel = ModelProviderFactory.buildChatModel(appConfig.llmProvider(), appConfig.apiKey());
Path outputDir = Paths.get(appConfig.absoluteOutputPath() + "/" + appConfig.projectName());
logger.info("🚀 Starting Tutorial Generation for: " + appConfig.inputPath());
try (TaskPipeline tasksPipeLine = new TaskPipeline()) {
tasksPipeLine.add("Code-crawler", new CodeCrawlerTask())
.then("Identify-abstraction", new IdentifyAbstractionsTask(chatModel, appConfig.projectName()))
.then("Identify-relationships", new IdentifyRelationshipsTask(chatModel, appConfig.projectName()))
.then("Chapter-organizer", new ChapterOrganizerTask(chatModel))
.then("Technical-writer", new TechnicalWriterTask(chatModel, outputDir))
.then("Meta-info", new MetaInfoTask(chatModel, outputDir, appConfig.projectName(), appConfig.inputPath()));
tasksPipeLine.run(appConfig.inputPath()).join();
logger.info("\n✅ Tutorial generation complete! Output located at: " + outputDir);
}
}
}
All of these Tasks in the TaskPipeline above are backed by a prompt. So basically its prompt chaining. There wasnt much about SKILLS.md back then. Fast track to today. I have tried to take all the prompts and put them in a Skill instead. And atleast from the intial tests I am looking at its looking pretty good. For example here is a mermaid drawing generated from one of Scribe's code base (an MCP server)
graph TD
KantraTool[KantraTool<br/>MCP Server Entry Point] -->|queries availability| CommandRegistry[CommandRegistry<br/>Plugin System]
KantraTool -->|invokes execute| KantraCommand[KantraCommand<br/>Command Interface]
KantraTool -->|accepts as parameter| KantraOperation[KantraOperation<br/>Operation Catalog]
CommandRegistry -->|discovers via CDI| KantraCommand
CommandRegistry -->|uses as map key| KantraOperation
CommandRegistry -->|applies filtering| CommandConfig[Configuration System]
KantraCommand -->|declares operation| KantraOperation
ConcreteCommands[Concrete Command<br/>Implementations] -->|extends| AbstractCommand[AbstractCommand<br/>Command Base Class]
ConcreteCommands -->|constructs| Rule[Rule<br/>Domain Model]
ConcreteCommands -->|creates instances| Condition[Condition<br/>Polymorphic Detection]
AbstractCommand -->|delegates serialization| RuleValidator[RuleValidator<br/>Validation Service]
AbstractCommand -->|converts to enums| DomainEnums[Domain Enumerations<br/>JavaLocation, Category]
AbstractCommand -->|constructs| SupportingModels[Supporting Model Classes<br/>Link, CustomVariable]
Rule -->|contains when field| Condition
Rule -->|uses category| DomainEnums
Rule -->|composes| SupportingModels
Condition -->|references| DomainEnums
Condition -->|contains| SupportingModels
RuleValidator -->|validates & serializes| Rule
RuleValidator -->|validates recursively| Condition
style KantraTool fill:#e1f5ff,stroke:#0066cc,stroke-width:3px
style Rule fill:#fff4e1,stroke:#ff9900,stroke-width:2px
style Condition fill:#fff4e1,stroke:#ff9900,stroke-width:2px
style CommandRegistry fill:#e8f5e9,stroke:#4caf50,stroke-width:2px
style RuleValidator fill:#e8f5e9,stroke:#4caf50,stroke-width:2px
Building the SKILLS.md
Lets take a look at the basic file structure when creating a skill.
File Structure
-
SKILL.md- Main skill definition and instructions -
package.json- NPM package configuration -
installer.js- Installation script -
bin/- Executable commands -
tests/- Test files, in my case these are test projects in multiple languages so I can compare whats generated.
The Skills file is quite large (100+ lines), here is a little snippet from the code discover phase during anaylsis.
1. Determine project details:
- If user provided path in command (e.g., `/tutorial analyze ./src`), use that
- Otherwise ask: "What directory should I analyze?"
- Auto-detect primary language from file extensions
- Ask if they want to focus on specific areas (optional)
2. Find source files using Glob:
- **Java**: `**/*.java` (exclude `**/test/**`, `**/target/**`)
- **Python**: `**/*.py` (exclude `**/test/**`, `**/__pycache__/**`, `**/venv/**`)
- **JavaScript/TypeScript**: `**/*.{js,ts,jsx,tsx}` (exclude `**/node_modules/**`, `**/dist/**`, `**/build/**`)
- **Go**: `**/*.go` (exclude `**/*_test.go`, `**/vendor/**`)
- **C#**: `**/*.cs` (exclude `**/bin/**`, `**/obj/**`)
- **Ruby**: `**/*.rb` (exclude `**/spec/**`, `**/test/**`)
- **Rust**: `**/*.rs` (exclude `**/target/**`)
- **PHP**: `**/*.php` (exclude `**/vendor/**`, `**/tests/**`)
3. Handle large codebases:
- If >50 files found, ask: "Found {N} files. Analyze all or focus on specific subdirectory?"
- Suggest core directories: `src/main`, `lib`, `app`, etc.
4. Read discovered files with Read tool
5. Report: "Found {N} files totaling {LOC} lines of code"
Quick start
npx @sshaaf/tutorial-skill install
# reload the coding agent. e.g. claude.
/tutorial build
#You can preview the files in markdown on disk. for example preview in VSCode.
#However if you want to preview in html then use the following command.
/tutorial preview
Skill Modes
Currently the Skill is built into two modes. This is exactly what Waver also did, analyze the code base and then generte a contextual tutorial. Here I have broken down that process, so if users just want very basic documentation like the mermaid diagram above with some markdown of the analysis they can do that. Else building tutorial which is the second mode has more steps, to make in consumable and nice 😎
/tutorial analyze - Analyze Mode
- 3-stage pipeline for quick codebase understanding
- Generates architecture diagrams with Mermaid
- NEW: Option to save diagrams to disk
- Time: 2-5 minutes
/tutorial build - Build Mode
- 6-stage pipeline for comprehensive tutorial generation
- Creates multi-chapter Markdown tutorials
- Time: 10-30 minutes
More Commands
While the above commands publish markdown onto disk, it can also be useful to preview them or even push them into github pages. Following will allow previewing locally and checking that it all works.
/tutorial preview
- Local tutorial preview with HonKit
- Time: 5-30 seconds to start
-
Output: Local docs site (usually
http://localhost:4000) - Use for: Reviewing generated docs before publishing
/tutorial doctor
- Diagnostics for local preview/runtime + docs scaffolding
- Time: ~10-30 seconds
- Output: Pass/fail checklist
-
Use for: Verifying HonKit runtime +
book.jsonbefore publishing
The markdown can also be pushed into github pages. Here is a nice example from @Kim Borup

How to use it: Tutorial Generator & Analyzer
# Quick analysis
/tutorial analyze .
# With path
/tutorial analyze ./src/main/java
# Full tutorial
/tutorial build .
# With output directory
/tutorial build --output ./docs/tutorial
# Preview generated tutorial in Claude mode
/tutorial preview ./docs/tutorial
# Diagnose local preview/runtime/docs scaffolding
/tutorial doctor ./docs/tutorial
# Initialize docs files for HonKit
npx @sshaaf/tutorial-skill docs init --dir ./docs/tutorial
# Preview locally with HonKit
npx @sshaaf/tutorial-skill docs preview --dir ./docs/tutorial
# Build static site with HonKit
npx @sshaaf/tutorial-skill docs build --dir ./docs/tutorial
# Diagnose runtime/plugin setup
npx @sshaaf/tutorial-skill docs doctor --dir ./docs/tutorial
The difference
SKILL works directly within Claude Code CLI - no separate tools needed, integrates into the developer workflow. No setup or installation and is pretty aware of the context.
Waver on the other hand complete control on which models are being used, can run parallel and join tasks with its use of JGraphlet, can easily integarte into CI and batch modes for multiple projects etc.
I am not entirely sure if there is an either or an or here, I would still prefer waver just because its a simplified CLI approach, which is pretty solid for multiple cases, however as soon as we speak about developer workflow a Skill makes a lot more sense when one is using Claude or OpenCode etc. I think there is another bit of optmization here that can be investigated where I can take the waver project and integrate it within the skill, so the skill can handover some of the work. A back an forth to take the best of the two.
I have to admit; defining constraints in plain English and letting a coding agent determine the implementation is quite daunting. It requires a significant mindset shift, and I’m eager to see how this scales and where the true limitations lie.
The skill is live, and the journey from a standalone cli to an NPM-distributable skill is complete. Now, the real testing begins 😎😊
View the project: github.com/sshaaf/tutorial-skill
Try it out: npx @sshaaf/tutorial-skill
Top comments (0)