Website | Get Started | Documentation | Twitter | Support
OneService allows you to run OneSDK as a service with configured usecases. And use it as backend for frontend.
For more details about Superface visit how it works and get started.
You can use this package as a globally installed CLI program:
npm install --global @superfaceai/one-service
To run OneService you need to have Superface configuration.
-
Create new folder where the configuration will be created:
mkdir myapp cd myapp
-
Install usecases and configure providers :
npx @superfaceai/cli install weather/current-city --providers wttr-in
(Repeat for any usecase you find in Catalog.)
-
Start OneService with GraphiQL.
oneservice --graphiql
-
Visit http://localhost:8000/ to open GraphQL interactive IDE
OneService package provides createGraphQLMiddleware
function for mounting the GraphQL server as a middleware with any HTTP web framework that supports connect styled middleware. This includes Connect itself, Express, Polka, Restify and others.
The createGraphQLMiddleware
function returns a promise, therefore you need to await resolution before mounting the middelware.
If you can use ES modules in your project, you can resolve the promise with top-level await
:
// server.mjs
import express from 'express';
import { createGraphQLMiddleware } from '@superfaceai/one-service';
const app = express();
const graphqlMiddleware = await createGraphQLMiddleware({
graphiql: true,
});
app.use('/graphql', graphqlMiddleware);
app.listen(3000);
Alternatively you can setup the server inside an async
funtion:
const express = require('express');
const { createGraphQLMiddleware } = require('@superfaceai/one-service');
async function startServer() {
const app = express();
const graphqlMiddleware = await createGraphQLMiddleware({
graphiql: true,
});
app.use('/graphql', graphqlMiddleware);
app.listen(3000);
}
startServer();
createGraphQLMiddleware
function will throw an error if it cannot generate a GraphQL schema. Possible issues include:
- missing or invalid
super.json
- name collision between use-cases of installed profiles
- profile with features unsupported by GraphQL
query WeatherInPrague {
WeatherCurrentCity {
GetCurrentWeatherInCity(input: { city: "Prague" }) {
result {
temperature
feelsLike
description
}
}
}
}
query SelectProvider {
WeatherCurrentCity {
GetCurrentWeatherInCity(
input: { city: "Prague" }
provider: { mock: { active: true } } # active flag is optional if only one provider is configured
) {
result {
temperature
feelsLike
description
}
}
}
}
query InstalledProfilesAndProviders {
_superJson {
profiles {
name
version
providers
}
providers
}
}
OneService doesn't automatically load .env
file. You can manually use dotenv package for this functionality. If you would like to use OneService CLI with dotenv, follow these steps:
-
Install dotenv and OneService locally in your project:
npm i dotenv @superfaceai/one-service
-
Run the CLI via
node
with dotenv preload:node -r dotenv/config node_modules/.bin/oneservice
OneService doesn't provide any authentication or CORS support. This should be handled by API Gateway (eg. Express Gateway or Kong) or you can use own Express instance and attach the middleware.
OneService can be deployed to Heroku with Git.
- Heroku CLI installation instructions
- Git installation instructions
- Node.js installation instructions
-
You will need folder for the application with local Git repository
mkdir myapp cd myapp git init
-
Next step is to install OneService
npm init -y npm install --save @superfaceai/one-service
-
Install profiles as explaied in Usage step 2.
-
echo 'web: oneservice --port $PORT --host 0.0.0.0 --graphiql --logs' > Procfile
-
Commit changes to Git repository
git add --all git commit -m 'OneService configuration'
-
Create Heroku remote
heroku create
-
Deploy app
git push heroku main
OneService can print structured logs. To allow it use --logs=<level>
flag:
oneservice --logs
oneservice --logs=trace
Set the optional logging level specifier. Available levels are (ordered by priority):
- fatal
- error
- warn
- info
- debug
- trace
- silent
The logging level is the minimum threshold. When set, every higher level will be logged as well. For instance if logger.level
is info
then all fatal
, error
, warn
, and info
logs will be enabled.
You can pass silent
to disable logging.
To print structured logs in readable way during testing or developement, you can use pino-pretty.
Install pino-pretty
:
npm install -g pino-pretty
Run OneService with pretty print:
oneservice --logs=trace | pino-pretty
We welcome all kinds of contributions! Please see the Contribution Guide to learn how to participate.
OneService is licensed under the MIT License.
© 2023 Superface s.r.o.