Your Cron Jobs Can't Think. These Can.
Every morning, the same ritual: open five tabs, skim headlines, paste into an LLM, wait for a summary, copy it to a file, forward it to Telegram. Manual. Repetitive. Skippable when you're busy β and that's exactly when you need it most.
I automated the entire thing β including the LLM call β in one TOML file. No Python. No Bash glue. No three separate cron entries. Zenii runs the whole pipeline on a schedule, passes outputs between steps, and fires off the Telegram message before I've poured my first coffee.
Here's exactly how it works.
The Workflow: Daily LLM Digest
Four steps. One schedule. The workflow fetches top news, summarises it with an LLM into a 5-bullet briefing, then fans out in parallel β saving the result to a local file and sending it to Telegram at the same time.
id = "daily-news-digest"
name = "Daily LLM Digest"
description = "Fetches top news, produces an LLM-summarized briefing, and sends to Telegram"
schema_version = 1
schedule = "15 11 * * *"
[[steps]]
name = "fetch_news"
type = "tool"
tool = "web_search"
args = { query = "top technology news today" }
[[steps]]
name = "summarize"
type = "llm"
prompt = "You are a news editor. Summarize the following search results into a concise 5-bullet briefing. Use markdown formatting.\n\n{{steps.fetch_news.output}}"
depends_on = ["fetch_news"]
[[steps]]
name = "save_briefing"
type = "tool"
tool = "file_write"
depends_on = ["summarize"]
args = { path = "/tmp/zenii/daily-briefing.md", content = "{{steps.summarize.output}}" }
[[steps]]
name = "notify_telegram"
type = "tool"
tool = "channel_send"
depends_on = ["summarize"]
args = { action = "send", channel = "telegram", message = "π° *Daily News Digest*\n\n{{steps.summarize.output}}" }
[layout]
fetch_news = { x = 100.0, y = 0.0 }
summarize = { x = 400.0, y = 0.0 }
save_briefing = { x = 700.0, y = 0.0 }
notify_telegram = { x = 700.0, y = 150.0 }
Enter fullscreen mode Exit fullscreen mode
How the Steps Connect
Zenii builds a directed acyclic graph (DAG) from the depends_on declarations. Cycles are rejected at save time. Steps with no dependencies run first; everything else waits for its upstream steps to complete.
In this workflow:
fetch_newsruns immediately β no dependenciessummarizewaits forfetch_news, then injects its output into the LLM prompt via{{steps.fetch_news.output}}save_briefingandnotify_telegramboth depend onsummarizeβ they execute in parallel, so the file write and the Telegram message happen at the same time
The [layout] section stores x/y coordinates for each node on the visual canvas. It has no effect on execution order β that's determined entirely by depends_on.
The template syntax {{steps.step_name.output}} works in any field β LLM prompts, tool args, condition expressions. You can also reference {{steps.step_name.success}} and {{steps.step_name.error}} for flow control.
For full DAG mechanics, retry policies, and failure handling, see Zenii Workflow Scheduling Documentation.
Create It in Plain English
You don't have to write the TOML by hand. Describe the workflow in plain English in Zenii's chat interface β it generates the full TOML, wires the depends_on graph, and sets the schedule for you.

Type what you want. Zenii generates the TOML, builds the graph, and registers the schedule.
The daily digest above was created from a single prompt:
"Create a daily news digest workflow that runs at 11:15 AM every day. It should search for top technology news, summarize the results, then save the briefing to /tmp/zenii/daily-briefing.md and send it to Telegram as 'Daily News Digest'."
The TOML in this post is exactly what Zenii produced. You can edit it afterward β or just describe a change and let the chat update it.
The Schedule Field
schedule = "15 11 * * *" registers the workflow with Zenii's built-in cron scheduler. Standard five-field cron syntax: minute hour day month weekday.
Expression
Meaning
15 11 * * *
Every day at 11:15 AM
0 9 * * 1-5
Weekdays at 9 AM
*/30 * * * *
Every 30 minutes
0 8 1 * *
First of every month at 8 AM
Zenii also supports interval syntax for simpler cases:
schedule = "every 300s" # run every 5 minutes
Enter fullscreen mode Exit fullscreen mode
And one-shot jobs that auto-delete after a successful run:
schedule = "0 10 * * *"
one_shot = true
Enter fullscreen mode Exit fullscreen mode
The scheduler persists across daemon restarts via SQLite. Missed runs are tracked, and failures retry with exponential backoff: 30s β 60s β 5m β 15m β 1h.
The Node Palette
The type = "tool" step type connects to every node Zenii ships. The type = "llm" step type is its own first-class citizen β it takes a prompt field and calls your configured AI provider directly.
Here's the full palette:
Category
Node
What it does
AI
llm_prompt
Run a prompt against your configured AI provider
Search
web_search
Search the web and return ranked results
Search
wiki_search
Query your Zenii wiki knowledge base
System
system_info
Read CPU, memory, and OS details
System
shell
Execute a shell command and capture output
System
process
Start, stop, or inspect OS processes
Files
file_read
Read a file's contents
Files
file_write
Write content to a file
Files
file_search
Search files by name or pattern
Files
file_list
List directory contents
Files
patch
Apply a unified diff patch to a file
Memory
memory_store
Write a key-value fact to long-term memory
Memory
memory_recall
Retrieve from memory by semantic query
Memory
memory_forget
Delete a memory entry by key
Channels
channel_send
Send a message to Telegram, Slack, or Discord
Config
config_read
Read a value from Zenii's config
Config
config_update
Update a config value at runtime
Flow Control
delay
Pause for N seconds (useful for rate limiting)
Flow Control
condition
Branch on a boolean expression
The wiki_search node is worth calling out: it queries your local Zenii wiki β your own indexed documents, notes, and saved knowledge β and returns relevant passages. Combined with an llm step, your scheduled workflow can synthesize current web results against your private knowledge base. See Stop Rereading Your Documents. Let the AI Study Them Once. for how to build and populate your wiki.
More Workflow Ideas
A few quick sketches using different parts of the palette.
Weekly code health check β run tests on a schedule, only alert on failure:
id = "weekly-test-check"
schedule = "0 9 * * 1"
[[steps]]
name = "run_tests"
type = "tool"
tool = "shell"
args = { command = "cargo test 2>&1" }
[[steps]]
name = "check_result"
type = "tool"
tool = "condition"
depends_on = ["run_tests"]
args = { expression = "{{steps.run_tests.success}}", if_false = "alert" }
[[steps]]
name = "alert"
type = "tool"
tool = "channel_send"
depends_on = ["check_result"]
args = { channel = "telegram", message = "Tests failed:\n\n{{steps.run_tests.output}}" }
Enter fullscreen mode Exit fullscreen mode
Memory-augmented research digest β search, summarize, and remember for next time:
id = "research-memory"
schedule = "0 18 * * *"
[[steps]]
name = "search"
type = "tool"
tool = "web_search"
args = { query = "Rust async runtime updates this week" }
[[steps]]
name = "summarize"
type = "llm"
prompt = "Summarize in 3 sentences:\n\n{{steps.search.output}}"
depends_on = ["search"]
[[steps]]
name = "remember"
type = "tool"
tool = "memory_store"
depends_on = ["summarize"]
args = { key = "rust-weekly-{{date}}", value = "{{steps.summarize.output}}" }
Enter fullscreen mode Exit fullscreen mode
Config-driven topic digest β change the search topic from config without touching the workflow:
id = "topic-digest"
schedule = "30 8 * * *"
[[steps]]
name = "get_topic"
type = "tool"
tool = "config_read"
args = { key = "digest.topic" }
[[steps]]
name = "search"
type = "tool"
tool = "web_search"
depends_on = ["get_topic"]
args = { query = "{{steps.get_topic.output}} news today" }
[[steps]]
name = "summarize"
type = "llm"
prompt = "Summarize the key points:\n\n{{steps.search.output}}"
depends_on = ["search"]
[[steps]]
name = "send"
type = "tool"
tool = "channel_send"
depends_on = ["summarize"]
args = { channel = "telegram", message = "{{steps.summarize.output}}" }
Enter fullscreen mode Exit fullscreen mode
Update the topic with zenii config set digest.topic "machine learning" and the next run picks it up.
Run It
Save the workflow file and register it:
zenii workflow create digest.toml
Enter fullscreen mode Exit fullscreen mode
Test it immediately without waiting for the cron:
zenii workflow run daily-news-digest
Enter fullscreen mode Exit fullscreen mode
Check the run history with per-step timing and output:
zenii workflow history daily-news-digest
Enter fullscreen mode Exit fullscreen mode
List all scheduled workflows and their next fire times:
zenii schedule list
Enter fullscreen mode Exit fullscreen mode
You can also trigger any workflow over HTTP from a CI pipeline, webhook, or MCP agent:
curl -X POST http://localhost:18981/workflows/daily-news-digest/run \
-H "Authorization: Bearer $TOKEN"
Enter fullscreen mode Exit fullscreen mode
One TOML file, one cron expression, one Zenii process. Your LLM pipeline runs while you sleep.
Full scheduling docs: https://docs.zenii.sprklai.com/scheduling
GitHub: https://github.com/sprklai/zenii β MIT licensed, open source.
If you build something with it, drop a link in the comments.

