Using rasa NLU as a HTTP server

Note

Before you can use the server, you need to train a model! See Training Your Model

The HTTP api exists to make it easy for non-python projects to use rasa NLU, and to make it trivial for projects currently using {wit,LUIS,api}.ai to try it out.

Running the server

You can run a simple http server that handles requests using your models with (single threaded)

$ python -m rasa_nlu.server -c config_spacy.json --server_model_dirs=./model_YYYYMMDD-HHMMSS

Emulation

rasa NLU can ‘emulate’ any of these three services by making the /parse endpoint compatible with your existing code. To activate this, either add 'emulate' : 'luis' to your config file or run the server with -e luis. For example, if you would normally send your text to be parsed to LUIS, you would make a GET request to

https://api.projectoxford.ai/luis/v2.0/apps/<app-id>?q=hello%20there

in luis emulation mode you can call rasa by just sending this request to

http://localhost:5000/parse?q=hello%20there

any extra query params are ignored by rasa, so you can safely send them along.

Endpoints

POST /parse (no emulation)

You must POST data in this format '{"q":"<your text to parse>"}', you can do this with

$ curl -XPOST localhost:5000/parse -d '{"q":"hello there"}'

POST /train

You can post your training data to this endpoint to train a new model. This request will wait for the server answer: either the model was trained successfully or the training errored. If you want to name your model to be able to use it during parse requests later on, you should pass the name /train?name=my_model. Any parameter passed with the query string will be treated as a configuration parameter of the model, hence you can change all the configuration values listed in the configuration section by passing in their name and the adjusted value.

$ curl -XPOST localhost:5000/train -d @data/examples/rasa/demo-rasa.json

GET /status

This checks if there is currently a training process running (you can only run one at a time). also returns a list of available models the server can use to fulfill /parse requests.

$ curl localhost:5000/status | python -mjson.tool
{
  "training" : False
  "models" : []
}

GET /version

This will return the current version of the Rasa NLU instance.

$ curl localhost:5000/version | python -mjson.tool
{
  "version" : "0.8.2"
}

GET /config

This will return the currently running configuration of the Rasa NLU instance.

$ curl localhost:5000/config | python -mjson.tool
{
    "config": "/app/rasa_shared/config_mitie.json",
    "data": "/app/rasa_nlu/data/examples/rasa/demo-rasa.json",
    "duckling_dimensions": null,
    "emulate": null,
    ...
  }

Authorization

To protect your server, you can specify a token in your rasa NLU configuration, e.g. by adding "token" : "12345" to your config file, or by setting the RASA_TOKEN environment variable. If set, this token must be passed as a query parameter in all requests, e.g. :

$ curl localhost:5000/status?token=12345

On default CORS (cross-origin resource sharing) calls are not allowed. If you want to call your rasa NLU server from another domain (for example from a training web UI) then you can whitelist that domain by adding it to the config value cors_origin.

Serving Multiple Apps

Depending on your choice of backend, rasa NLU can use quite a lot of memory. So if you are serving multiple models in production, you want to serve these from the same process & avoid duplicating the memory load.

Although this saves the backend from loading the same backend twice, it still needs to load one set of
word vectors (which make up most of the memory consumption) per language and backend.

You can use the multi-tenancy mode by replacing the server_model_dirs config variable with a json object describing the different models.

For example, if you have a restaurant bot and a hotel bot, your configuration might look like this:

{
  "server_model_dirs": {
    "hotels" : "./model_XXXXXXX",
    "restaurants" : "./model_YYYYYYY"
  }
}

You then pass an extra model parameter in your calls to /parse to specify which one to use:

$ curl 'localhost:5000/parse?q=hello&model=hotels'

or

$ curl -XPOST localhost:5000/parse -d '{"q":"I am looking for Chinese food", "model": "restaurants"}'

If one of the models is named default, it will be used to serve requests missing a model parameter. If no model is named default requests without a model parameter will be rejected.