Skip to main content
Selora Homes Selora Homes
Product Updates

Selora AI v0.8.0: Selora AI Cloud, Token Tracking & Savings, and Scenes Everywhere

Selora AI v0.8.0 ships our managed Selora AI Cloud provider, a new Usage panel that tracks token consumption with cost estimates, optimizations that cut token usage on the hourly suggestion job, end-to-end Scenes support, and a redesigned chat with native Home Assistant device cards.

Selora AI v0.8.0: Selora AI Cloud, Token Tracking & Savings, and Scenes Everywhere

Selora AI v0.8.0 is out. This release closes the loop on three things we’ve been building toward for months: a managed cloud backend so you don’t have to bring your own keys, full visibility into where your tokens (and dollars) go plus optimizations that meaningfully cut how many tokens we burn in the first place, and a Scenes experience that finally feels native in chat.


Selora AI Cloud Provider

The biggest shift in v0.8.0 is the new Selora AI Cloud provider – a managed LLM backend routed through the Selora AI Gateway. Sign in with your Selora account via OAuth (PKCE), and you’re done. No API keys to paste, no provider to pick, no model routing to configure. Multi-model selection happens server-side, so you always get the right model for the job without thinking about it.

If you’re a Selora Hub customer, your Cloud tokens are now provisioned automatically and consumed on first launch – the integration picks them up without any manual setup.

For everyone else, free Selora accounts get 5,000 Cloud credits (a one-time allotment) so you can try the managed provider without setting up any third-party API key. Paid plans for users without a Selora Hub subscription – both a monthly subscription and pay-as-you-go top-ups – are landing soon for when you outgrow the free credits.

The direct providers (Anthropic, OpenAI, Gemini, OpenRouter, Ollama) all remain fully supported if you’d rather bring your own keys.


Token Tracking and Smart Savings

We tackled tokens from two directions in v0.8.0: showing you exactly where they go, and using fewer of them in the first place.

Track every token

A new Usage tab gives you a real-time and historical view of token consumption across every conversation, suggestion run, and snapshot. You’ll see input vs. output tokens, cached tokens, breakdowns by provider and model, and cost estimates in dollars so there are no surprises at the end of the month. This is the same telemetry we use internally to tune the integration, exposed directly in the panel.

Use fewer tokens

Alongside the new visibility, v0.8.0 also makes the integration cheaper to run. The collector and suggestion generator now reuse cached enrichment data, deduplicate snapshots, and skip redundant LLM calls. Typical homes see substantial reductions on the hourly suggestion job without any change in suggestion quality. Combined with the new Usage panel, you can watch the savings land in real time.


Scenes, End-to-End

The Scenes feature introduced in v0.7.0 is now fully integrated across the product:

  • The Scenes tab lists all of your Home Assistant scenes, not just the ones Selora AI created
  • Scenes are exposed via MCP, so the LLM can list and activate them directly in chat
  • Inline scene cards show up in conversations with one-tap activate and edit actions
  • Mobile alignment fixes for the activate and burger buttons

You can now say “activate movie night” in chat and have it run, no matter how the scene was originally created.


Composer & Header Redesign

The chat composer and panel header have been refreshed:

  • Quick actions surfaced in the composer
  • Streamlined header with less visual noise
  • Keyboard shortcuts: Enter to send, Shift+Enter for newline, Cmd/Ctrl+K to focus the composer

Small details, but they add up to a much faster feel in day-to-day use.


Native HA Device Cards in Conversations

When the LLM references a device in chat, Selora AI now embeds Home Assistant’s real entity card inline – with native controls. Toggle a switch, drag a volume slider, read a climate entity, control covers with arrow buttons – all without leaving the conversation.

This is the first step toward making the chat a complete control surface for your home, not just a way to ask the LLM to do things.

Asking Selora AI which lights are on, with native HA device cards rendered inline


Better Failure UX

Transient failures used to be silent or cryptic. v0.8.0 cleans that up:

  • LLM 429 / quota errors now surface as a clear red-particle alert instead of failing silently
  • Stream cutoffs are detected and reported with the real upstream error, not a generic timeout
  • WebSocket auto-reconnect on transient drops, so a flaky network no longer kills your session

Improvements

  • Faster provider health checks during config (Anthropic, Gemini, Ollama, and OpenRouter run in parallel with shorter timeouts)
  • Background service calls in the panel, so the UI no longer blocks on slow Home Assistant reloads
  • Hub-provisioned Selora Cloud tokens are consumed automatically on first launch
  • select_devices config flow now surfaces real errors instead of a generic “abort”

Bug Fixes

  • Narrow-panel button visibility and overflow on mobile
  • Entity chips and grid padding flicker on tab switch
  • Options flow TypeError when clicking the config gear
  • Black screen on panel re-entry
  • Enrichment cache sync and migration edge cases
  • Dev-mode URL override for local panel development

See the full changelog for the complete list.


What’s Next

Our next release is focused on Selora AI Local : an in-house model that runs entirely on your own hardware – either on the same device as Home Assistant or on any Ollama server you point it at. Selora Hub customers will get the usual one-click install to flip it on.

Local Selora AI means no data leaving the home network, no token costs, and no dependency on an external provider for the day-to-day work the assistant does. We’ll have more to share soon.

Update via HACS, or if you already have Selora AI installed, the update will appear in your Home Assistant settings.

Type to search across cities, counties, and installers

↑↓ navigate open
⌘K search