Skip to content
This site is under construction. Content, screenshots, and workflows may change at any time.

Sources

The Knowledge Base page is the bot’s source manager. It’s where you add new sources, monitor indexing status, retry failed ones, and tell the crawler which URLs to skip.

For the first-time setup walkthrough, see Train your bot. This page covers ongoing management.

Knowledge Base view with All Sources tab — table showing two source rows (Indexed and Failed), Add Source button, filter and sort controls

Tabs

  • All Sources — every source attached to this bot.
  • Blocked URLs — paths the crawler should skip across all Website Links sources for this bot.

Sources table

Columns:

ColumnNotes
TypeOne of URL, URL List, Crawl, Sitemap, File, Text, Fluent Support.
DescriptionSource title or origin URL / filename.
DocumentsIndexed-document count.
StatusBadge — see status table below.

Toolbar

Above the table:

  • Source type dropdownAll source types or one specific type.
  • Search icon — filter rows by title / description.
  • Filter — additional filters (e.g. status).
  • Sort bycreated_at, title, or documents_count.
  • + Add Source — opens the source picker (see Train your bot).

Status states

StatusMeaning
QueuedWaiting to start processing.
ScrapingFetching content from the source.
IndexingSaving extracted documents and making them searchable.
IndexedReady to answer questions.
Partially IndexedSome items succeeded, some failed. Open and click Retry.
FailedEvery item failed or the source itself is unreachable. Open and click Retry.
DeletingRemoval in progress.

Status and document counts update automatically while processing runs.

Source dialog

Click any row to open the source dialog. Contents depend on type:

  • Status badge + last update timestamp.
  • Failure reason (if Failed).
  • Retry button — only shown when status is Failed or Partially Indexed.
  • Documents list — with per-document Delete document action (no per-document retry).
  • Scheduled Refresh dropdown — Never / Daily / Weekly / Monthly. Website Links sources only.
  • Delete Source — permanent removal.

See Retrain & update for refresh and retry semantics.

Blocked URLs

A per-bot deny-list. The crawler skips matching URLs for every Website Links source on this bot.

Blocked URLs tab with empty state and 'Add URL to block' button

Click Add URL to block to open the dialog:

  • URL field — exact URL to block (e.g. https://example.com/careers).
  • Also ignore all subpaths toggle — when on, every path under the URL is also skipped (e.g. blocking /blog ignores /blog/post-1).
Add URL to block dialog with URL field, Block button, and 'Also ignore all subpaths' toggle

Existing blocked URLs show in the table with a Remove action.

Use this for paths you never want the bot to learn from: careers pages, legal boilerplate, log-in walls, marketing-only sections that confuse answers.

Troubleshooting

  • Status stuck on Scraping — check the source dialog for a failure reason. The source may be hitting rate limits or auth challenges.
  • Same URL keeps failing — likely a 404 or login wall. Add it to Blocked URLs so it’s skipped on future refreshes.
  • Documents count dropped after refresh — pages that 404 since the last crawl are removed. Open the source dialog to see what disappeared.

What’s next