kordelfrance's picture
Upload folder using huggingface_hub
8415b2f verified
# Scentience Skills
Agent skill definitions for olfaction-aware AI — part of the [Scentience](https://scentience.ai)
open olfaction platform. Built to the [agentskills.io](https://agentskills.io) open standard;
compatible with Claude Code, OpenAI Codex, and Google Antigravity.
## Skill Registry
| Skill | Purpose |
|-------|---------|
| [ble-device](./ble-device/SKILL.md) | Connect to Reconnaisscent or Scentinel via BLE; sample or stream raw sensor readings |
| [olfactory-navigation](./olfactory-navigation/SKILL.md) | Plume contact classification and chemical source localization |
| [olfactory-inertial-odometry](./olfactory-inertial-odometry/SKILL.md) | GPS-denied position estimation by fusing olfaction signals with IMU |
| [colip-embeddings](./colip-embeddings/SKILL.md) | Cross-modal retrieval and semantic labeling via COLIP models |
## Decision Guide
| Intent | Use |
|--------|-----|
| Connect to hardware, read sensor data | [ble-device](./ble-device/SKILL.md) |
| I have live OPU readings — where should I go? | [olfactory-navigation](./olfactory-navigation/SKILL.md) |
| Navigate without GPS using smell + motion | [olfactory-inertial-odometry](./olfactory-inertial-odometry/SKILL.md) |
| Semantically label or match a smell episode | [colip-embeddings](./colip-embeddings/SKILL.md) |
| Classify a chemical compound from sensor data | [colip-embeddings](./colip-embeddings/SKILL.md) (`ovl-classifier` model) |
## Typical Workflows
**Real-time robotic navigation:**
`ble-device``olfactory-navigation` → motor controller
**GPS-denied localization:**
`ble-device``olfactory-inertial-odometry` → path planner
**Semantic labeling pipeline:**
`ble-device``colip-embeddings` → natural language annotation
**Full hybrid system:**
`ble-device``olfactory-navigation` + `olfactory-inertial-odometry` + `colip-embeddings`
## Design Principles
- **Olfaction as first-class modality** — chemical sensing receives the same structural
rigor as vision or audio; no scalar collapsing of temporal signals
- **Temporal fidelity** — all skills preserve signal history, intermittency, and trend
context across the full reading window
- **Calibrated uncertainty** — every output carries a `confidence` field; no skill
returns conclusions without explicit bounds
- **Control ≠ semantics** — navigation decisions (`olfactory-navigation`,
`olfactory-inertial-odometry`) and semantic interpretation (`colip-embeddings`) are
deliberately separate skills with separate output schemas
- **SDK-grounded** — all examples use the official Scentience SDK; no pseudocode
## SDK
| Language | Package | Install |
|----------|---------|---------|
| Python | [scentience](https://pypi.org/project/scentience/) | `pip install scentience` |
| JavaScript | [scentience](https://npmjs.com/package/scentience) | `npm install scentience` |
| Rust | [scentience](https://crates.io/crates/scentience) | `cargo add scentience` |
| C++ | [scentience/2.0.0](https://scentience.jfrog.io/ui/packages/conan:%2F%2Fscentience/2.0.0) | Conan |
API docs: [scentience.github.io/docs-api](https://scentience.github.io/docs-api/) ·
API keys: [dashboard.scentience.ai](https://dashboard.scentience.ai)