Warning: This document is for an old version of rasa NLU.

Tutorial: A simple restaurant search bot

Note

See Migrating an existing app for how to clone your existing wit/LUIS/api.ai app.

As an example we’ll use the domain of searching for restaurants. We’ll start with an extremely simple model of those conversations. You can build up from there.

Let’s assume that anything our bot’s users say can be categorized into one of the following intents:

  • greet
  • restaurant_search
  • thankyou

Of course there are many ways our users might greet our bot:

  • Hi!
  • Hey there!
  • Hello again :)

And even more ways to say that you want to look for restaurants:

  • Do you know any good pizza places?
  • I’m in the North of town and I want chinese food
  • I’m hungry

The first job of rasa NLU is to assign any given sentence to one of the categories: greet, restaurant_search, or thankyou.

The second job is to label words like “Mexican” and “center” as cuisine and location entities, respectively. In this tutorial we’ll build a model which does exactly that.

Preparing the Training Data

The best way to get training data is from real users, and the best way to do that is to pretend to be the bot yourself. But to help get you started we have some data saved here

Download the file and open it, and you’ll see a list of training examples like these:

{
  "text": "hey",
  "intent": "greet",
  "entities": []
}
{
  "text": "show me chinese restaurants",
  "intent": "restaurant_search",
  "entities": [
    {
      "start": 8,
      "end": 15,
      "value": "chinese",
      "entity": "cuisine"
    }
  ]
}

hopefully the format is intuitive if you’ve read this far into the tutorial, for details see Training Data Format.

In your working directory, create a data folder, and copy the demo-rasa.json file there.

Visualizing the Training Data

It’s always a good idea to look at your data before, during, and after training a model. There’s a great tool for creating training data in rasa’s format here - created by @azazdeaz - and it’s also extremely helpful for inspecting existing data.

For the demo data the output should look like this:

../_images/rasa_nlu_intent_gui.png

It is strongly recommended that you view your training data in the GUI before training.

Training Your Model

Now we’re going to create a configuration file. Make sure first that you’ve set up a backend, see Installation . Create a file called config.json in your working directory which looks like this

{
  "backend": "spacy_sklearn",
  "path" : "./models",
  "data" : "./data/examples/rasa/demo-rasa.json"
}

or if you’ve installed the MITIE backend instead:

{
  "backend": "mitie",
  "mitie_file": "./data/total_word_feature_extractor.dat",
  "path" : "./models",
  "data" : "./data/examples/rasa/demo-rasa.json"
}

Now we can train a spacy model by running:

$ python -m rasa_nlu.train -c config_spacy.json

After a few minutes, rasa NLU will finish training, and you’ll see a new dir called something like models/model_YYYYMMDD-HHMMSS with the timestamp when training finished.

Using Your Model

To run your trained model, pass the configuration value server_model_dirs when running the server:

$ python -m rasa_nlu.server -c config_spacy.json --server_model_dirs=./model_YYYYMMDD-HHMMSS

The passed model path is relative to the path configured in the configuration. More information about starting the server can be found in Using rasa NLU as a HTTP server.

You can then test our your new model by sending a request. Open a new tab/window on your terminal and run

$ curl -XPOST localhost:5000/parse -d '{"q":"I am looking for Chinese food"}' | python -mjson.tool

which should return

{
  "intent" : "restaurant_search",
  "confidence": 0.6127775465094253,
  "entities" : [
    {
      "start": 8,
      "end": 15,
      "value": "chinese",
      "entity": "cuisine"
    }
  ]
}

If you are using the spacy_sklearn backend and the entities aren’t found, don’t panic! This tutorial is just a toy example, with far too little training data to expect good performance. rasa NLU will also print a confidence value. You can use this to do some error handling in your bot (maybe asking the user again if the confidence is low) and it’s also helpful for prioritising which intents need more training data.

With very little data, rasa NLU can in certain cases already generalise concepts, for example:

$ curl -XPOST localhost:5000/parse -d '{"q":"I want some italian"}' | python -mjson.tool
{
  "entities": [
    {
      "end": 19,
      "entity": "cuisine",
      "start": 12,
      "value": "italian"
    }
  ],
  "intent": "restaurant_search",
  "text": "I want some italian"
  "confidence": 0.4794813722432127
}

even though there’s nothing quite like this sentence in the examples used to train the model. To build a more robust app you will obviously want to use a lot more data, so go and collect it!