Configuration Guide
Horizon is configured through two files: a .env file for API keys and a data/config.json file for sources, AI provider, and filtering options.
AI Providers
Configure which AI model scores and summarizes your content.
Anthropic Claude:
{
"ai": {
"provider": "anthropic",
"model": "claude-sonnet-4.5-20250929",
"api_key_env": "ANTHROPIC_API_KEY",
"throttle_sec": 0
}
}
OpenAI:
{
"ai": {
"provider": "openai",
"model": "gpt-4",
"api_key_env": "OPENAI_API_KEY",
"throttle_sec": 0
}
}
MiniMax:
{
"ai": {
"provider": "minimax",
"model": "MiniMax-M2.7",
"api_key_env": "MINIMAX_API_KEY",
"throttle_sec": 0
}
}
Available models: MiniMax-M2.7, MiniMax-M2.7-highspeed, MiniMax-M2.5, MiniMax-M2.5-highspeed
Aliyun DashScope (OpenAI-compatible):
{
"ai": {
"provider": "ali",
"model": "qwen-plus",
"api_key_env": "DASHSCOPE_API_KEY",
"throttle_sec": 0
}
}
Use the DashScope compatible-mode endpoint. Set DASHSCOPE_API_KEY in your .env. Optional: set base_url to override the default https://dashscope.aliyuncs.com/compatible-mode/v1.
AI throttling
If your model has a strict per-minute request cap, you can slow the scorer down in data/config.json:
{
"ai": {
"throttle_sec": 4.5
}
}
throttle_sec: Pause between scored items in seconds. Default is0.4.5is a reasonable starting point for free-tier models capped around 15 requests per minute.- Set it back to
0if you have enough throughput headroom and want maximum speed.
Custom Base URL (for proxies):
{
"ai": {
"provider": "anthropic",
"base_url": "https://your-proxy.com/v1",
...
}
}
Information Sources
All sources are configured under the top-level sources key in config.json.
GitHub
{
"sources": {
"github": [
{
"type": "user_events",
"username": "gvanrossum",
"enabled": true
},
{
"type": "repo_releases",
"owner": "python",
"repo": "cpython",
"enabled": true
}
]
}
}
Hacker News
{
"sources": {
"hackernews": {
"enabled": true,
"fetch_top_stories": 30,
"min_score": 100
}
}
}
RSS Feeds
{
"sources": {
"rss": [
{
"name": "Blog Name",
"url": "https://example.com/feed.xml",
"enabled": true,
"category": "ai-ml"
}
]
}
}
{
"sources": {
"reddit": {
"enabled": true,
"fetch_comments": 5,
"subreddits": [
{
"subreddit": "MachineLearning",
"sort": "hot",
"fetch_limit": 25,
"min_score": 10
}
],
"users": [
{
"username": "spez",
"sort": "new",
"fetch_limit": 10
}
]
}
}
}
Requires an Apify account. Set APIFY_TOKEN in your .env file. The free tier includes $5/month of credit, enough for roughly 20,000 tweets.
{
"sources": {
"twitter": {
"enabled": true,
"users": ["karpathy", "ylecun"],
"fetch_limit": 10,
"fetch_reply_text": false,
"max_replies_per_tweet": 3,
"max_tweets_to_expand": 10,
"reply_min_likes": 5
}
}
}
users— Twitter screen names to monitor, without the@prefixfetch_limit— maximum tweets to fetch per run (across all users combined; minimum 100 due to actor constraint)fetch_reply_text— whentrue, fetch actual reply bodies for important tweets and append them under--- Top Comments ---so the AI can factor in community discussion. Disabled by default.max_replies_per_tweet— maximum reply lines to append per tweet (default: 3)max_tweets_to_expand— cap on how many tweets get reply expansion per run, to control Apify credit usage (default: 10)reply_min_likes— only include replies with at least this many likes (default: 0)
The scraper uses the altimis/scweet actor by default. You can override it with actor_id if needed.
Filtering
Content is scored 0-10:
- 9-10: Groundbreaking - Major breakthroughs, paradigm shifts
- 7-8: High Value - Important developments, deep technical content
- 5-6: Interesting - Worth knowing but not urgent
- 3-4: Low Priority - Generic or routine content
- 0-2: Noise - Spam, off-topic, or trivial
{
"filtering": {
"ai_score_threshold": 7.0,
"time_window_hours": 24
}
}
ai_score_threshold: Only include content scoring >= this valuetime_window_hours: Fetch content from last N hours
Environment Variable Substitution
RSS feed URLs support ${VAR_NAME} syntax for secrets. The variable is expanded at runtime from environment variables (or .env file):
{
"name": "LWN.net",
"url": "https://lwn.net/headlines/full_text?key=${LWN_KEY}",
"enabled": true
}
This way config.json can be committed to a public repo without leaking tokens.
Email Subscription
Email delivery is optional and disabled unless email.enabled is true. Horizon uses SMTP to send daily summaries and IMAP to check subscribe/unsubscribe requests.
{
"email": {
"enabled": true,
"smtp_server": "smtp.qq.com",
"smtp_port": 465,
"imap_server": "imap.qq.com",
"imap_port": 993,
"email_address": "xxx@qq.com",
"password_env": "EMAIL_PASSWORD",
"sender_name": "Horizon Daily",
"subscribe_keyword": "SUBSCRIBE",
"unsubscribe_keyword": "UNSUBSCRIBE"
}
}
enabled: Turns email subscription handling and daily email delivery on or off.smtp_server/smtp_port: SMTP server used to send emails.imap_server/imap_port: IMAP server used to scan incoming subscription requests.email_address: Sender account and mailbox checked for subscription requests.password_env: Environment variable containing the email password or app password. Defaults toEMAIL_PASSWORD.sender_name: Display name shown in sent emails.subscribe_keyword/unsubscribe_keyword: Keywords Horizon looks for in incoming email subjects.
Webhook Notification
Webhook notification is optional and disabled unless webhook.enabled is true. Horizon can call Feishu/Lark, DingTalk, Slack, Discord, or any custom webhook endpoint when the pipeline succeeds or fails.
{
"webhook": {
"enabled": true,
"url_env": "HORIZON_WEBHOOK_URL",
"delivery": "summary",
"overview_position": "first",
"platform": "generic",
"layout": "markdown",
"fallback_layout": "markdown",
"languages": null,
"request_body": {
"text": "#{message_title}\n#{summary}"
},
"headers": ""
}
}
enabled: Turns webhook delivery on or off. The default isfalse.url_env: Environment variable that contains the webhook URL. For example, setHORIZON_WEBHOOK_URL=https://...in.env.delivery: Controls how messages are sent. Usesummaryfor one full message, orsummary_and_itemsfor one overview message followed by one message per selected item.overview_position: Controls where the overview is sent insummary_and_itemsmode. Usefirstfor the traditional order, orlastto send item details in reverse and keep the overview as the newest chat message.platform: Optional webhook platform hint. Usegenericby default, orfeishu/larkto enable platform-specific card rendering.layout: Controls the message layout. Usemarkdownfor templated Markdown delivery, orcollapsiblewithplatform: "feishu"/"lark"for a single Feishu Card JSON 2.0 message with each item in a collapsed panel.fallback_layout: Reserved fallback layout for unsupported platform/layout combinations. The current safe fallback ismarkdown.languages: Optional webhook-only language filter. Use["zh"]or["en"]to send only selected languages; usenullor omit it to send all configuredai.languages.request_body: Optional request body. If empty, Horizon sends aGETrequest. If provided, Horizon sends aPOSTrequest.headers: Optional custom headers, oneKey: Valuepair per line.
When request_body is a JSON object or array, Horizon renders placeholders and serializes it as JSON. When it is a string, Horizon renders it directly and detects JSON if the rendered string is valid JSON.
Webhook Templates
Available variables:
| Variable | Description |
|---|---|
#{date} |
Report date, for example 2026-04-24 |
#{language} |
Language code, such as en or zh |
#{important_items} |
Number of items that passed the score threshold |
#{all_items} |
Total number of fetched items |
#{result} |
success or failed |
#{timestamp} |
Unix timestamp |
#{message_title} |
Message title, such as the daily title, overview title, or item title |
#{message_kind} |
Message kind: summary, overview, item, failure, or manual |
#{summary} |
Message Markdown. In summary_and_items mode this is the overview or one item body, depending on the message |
When delivery is summary_and_items, item messages also include:
| Variable | Description |
|---|---|
#{item_index} |
1-based item number |
#{item_count} |
Total number of item messages |
#{item_title} |
Current item title |
#{item_url} |
Current item URL |
#{item_score} |
Current item AI score |
For webhook delivery, Horizon flattens HTML disclosure blocks such as <details><summary>...</summary> in #{summary} into plain Markdown link lists. This makes the generated summary easier to render in chat products. Saved Markdown files, GitHub Pages, and email content are unchanged.
Use #{key?limit=N&split=DELIM} to truncate long values by splitting on DELIM and keeping segments until the total character count reaches N.
#{summary?limit=3000&split=---}
DingTalk
In DingTalk, create a custom group robot and use a custom keyword such as Horizon. The keyword must appear in the body content.
{
"msgtype": "markdown",
"markdown": {
"title": "Horizon #{date} Daily",
"text": "Horizon result: #{result}\n\nHorizon important items: #{important_items}/#{all_items}\n\n#{summary}"
}
}
Feishu / Lark
In Feishu or Lark, create a custom group robot and use a custom keyword such as Horizon. The keyword must appear in the body content.
Use Card JSON 2.0 for Markdown rendering. The card must include "schema": "2.0" and put rich-text Markdown components under card.body.elements.
To keep the group chat compact while still allowing readers to browse the full briefing inside Feishu, use the collapsible layout:
{
"webhook": {
"enabled": true,
"url_env": "HORIZON_WEBHOOK_URL",
"platform": "feishu",
"layout": "collapsible",
"fallback_layout": "markdown",
"languages": ["zh"]
}
}
With this layout, Horizon sends one interactive card containing the overview and one collapsed panel per selected item. Each panel can be expanded in Feishu to read the full item detail. The regular request_body template is ignored for this rendered card.
{
"msg_type": "interactive",
"card": {
"schema": "2.0",
"config": {
"wide_screen_mode": true
},
"header": {
"title": {
"tag": "plain_text",
"content": "#{message_title}"
},
"template": "blue"
},
"body": {
"elements": [
{
"tag": "markdown",
"content": "Horizon result: #{result}\nHorizon important items: #{important_items}/#{all_items}"
},
{
"tag": "hr"
},
{
"tag": "markdown",
"content": "#{summary}"
}
]
}
}
}
Static Site
Horizon writes generated summaries to data/summaries/ and copies publishable Markdown into docs/ for the GitHub Pages site. The repository includes a ready-to-use workflow at .github/workflows/daily-summary.yml.
To use GitHub Pages, enable Pages for the repository and run the scheduled workflow or trigger it manually. The generated site is built from the docs/ directory.
MCP Server
Horizon includes an MCP server for AI assistants and MCP-compatible clients.
uv run horizon-mcp
Available tools include hz_validate_config, hz_fetch_items, hz_score_items, hz_filter_items, hz_enrich_items, hz_generate_summary, and hz_run_pipeline.
See src/mcp/README.md for the full tool reference and src/mcp/integration.md for client setup.