OpenAI API
Use this when you are targeting a standard OpenAI-compatible /v1 backend. It is the broadest preset for cloud or self-hosted services that follow the OpenAI request shape.
Remote endpoints let Noema connect to inference servers that live on desktops, lab machines, or cloud VMs while you keep the same library and chat flow on your Apple device.
Want your phone to use the model running on your Mac without manual backend setup? Turn on Noema Relay and let devices discover each other automatically. Read the quick guide in Noema Relay.
Good for OpenAI-compatible /v1 backends.
Good when you want a large hosted catalog and provider routing.
Good for LM Studio’s local or hosted REST server.
Good for simple local or LAN-hosted Ollama usage.
Good for pairing another device with your Mac and using the provider configured there.
Remote registration depends on current backend presets, authentication flow, and catalog handling. Make sure the Noema device you are using and the machine hosting the backend are both on recent builds before troubleshooting connectivity.
Authorization: Bearer ....Use OpenRouter when you want access to OpenRouter's hosted model catalog and provider network instead of connecting to a single local server.
https://openrouter.ai. Noema fills this in automatically.OpenRouter is the best remote option when you want a large hosted model catalog, provider routing, and per-model capability-aware settings without manually wiring a separate inference server.
Use this when you are targeting a standard OpenAI-compatible /v1 backend. It is the broadest preset for cloud or self-hosted services that follow the OpenAI request shape.
Large hosted model catalog with provider routing.
Use OpenRouter when you want a broad hosted catalog and provider routing without manually wiring separate endpoints for every hosted model.
Per-model remote settings are available, including context length and standard sampling controls like temperature, top-p, top-k, min-p, and repetition penalty. Models can also surface provider defaults.
Connect to LM Studio’s local or hosted REST server.
Use LM Studio when you want Noema to connect to a machine that already serves models over LM Studio’s REST interface.
LM Studio models expose a settings sheet with context length plus standard sampling controls such as temperature, top-p, top-k, min-p, and repetition penalty.
Use Ollama when you want simple local or LAN-hosted model serving and easy model pulls. Expose the server properly if you need Noema to reach it from another device.
Use Noema Relay when you want to pair another device with your Mac and use the provider configured there. It is the least manual route when you are already inside the Noema ecosystem.
Provider presets help fill chat and catalog routes for OpenAI API, OpenRouter, LM Studio, and Ollama, but you can still override fields with absolute URLs or relative paths. Use the authentication drawer for headers such as Authorization when the provider requires them.
After you save, Noema fetches the remote catalog and records the HTTP outcome in a connection summary. Status chips in the list and detail views mirror that state, and a manual reload control lets you refresh on demand. When Off-Grid mode is active, network calls pause until you reconnect.
Remote per-model settings are currently centered on LM Studio and OpenRouter. In both cases, Noema can expose context length and standard sampling controls like temperature, top-p, top-k, min-p, and repetition penalty.
Add custom identifiers whenever a server omits a deployment you still want to address. Noema deduplicates them against fetched entries and folds them back into the picker so one-off builds remain selectable.