Content determination

From Wikipedia, the free encyclopedia
Jump to: navigation, search

Content determination is the subtask of Natural language generation that involves deciding on the information to be communicated in a generated text. It is closely related to the task of Document structuring.


Consider an NLG system which summarises information about sick babies.[1] Suppose this system has four pieces of information it can communicate

  1. The baby is being given morphine via an IV drop
  2. The baby's heart rate shows bradycardia's (temporary drops)
  3. The baby's temperature is normal
  4. The baby is crying

Which of these bits of information should be included in the generated texts?


There are three general issues which almost always impact the content determination task, and can be illustrated with the above example.

Perhaps the most fundamental issue is the communicative goal of the text, i.e. its purpose and reader. In the above example, for instance, a doctor who wants to make a decision about medical treatment would probably be most interested in the heart rate bradycardias, while a parent who wanted to know how her child was doing would probably be more interested in the fact that the baby was being given morphine and was crying.

The second issue is the size and level of detail of the generated text. For instance, a short summary which was sent to a doctor as a 160 character SMS text message might only mention the heart rate bradycarias, while a longer summary which was printed out as a multipage document might also mention the fact that the baby is on a morphine IV.

The final issue is how unusual and unexpected the information is. For example, neither doctors nor parents would place a high priority on being told that the baby's temperature was normal, if they expected this to be the case.

Regardless, content determination is very important to users, indeed in many cases the quality of content determination is the most important factor (from the user's perspective) in determining the overall quality of the generated text.


There are three basic approaches to document structuring: schemas (content templates), statistical approaches, and explicit reasoning.

Schemas [2] are templates which explicitly specify the content of a generated text (as well as Document structuring information). Typically they are constructed by manually analysing a corpus of human-written texts in the target genre, and extracting a content template from these texts. Schemas work well in practice in domains where content is somewhat standardised, but work less well in domains where content is more fluid (such as the medical example above).

Statistical techniques use statistical corpus analysis techniques to automatically determine the content of the generated texts. Such work is in its infancy, and has mostly been applied to contexts where the communicative goal, reader, size, and level of detail are fixed. For example, generation of newswire summaries of sporting events.[3][4]

Explicit reasoning approaches have probably attracted the most attention from researchers. The basic idea is to use AI reasoning techniques (such as knowledge-based rules,[1] planning,[5] pattern detection,[6] case-based reasoning,[7] etc.) to examine the information available to be communicated (including how unusual/unexpected it is), the communicative goal and reader, and the characteristics of the generated text (including target size), and decide on the optimal content for the generated text. A very wide range of techniques has been explored, but there is no consensus as to which is most effective.


  1. ^ a b Portet F, Reiter E, Gatt A, Hunter J, Sripada S, Freer Y, Sykes C (2009). "Automatic Generation of Textual Summaries from Neonatal Intensive Care Data". Artificial Intelligence. 173: 789–816. doi:10.1016/j.artint.2008.12.002. 
  2. ^ K McKeown (1985). Text Generation. Cambridge University Press
  3. ^ R Barzilay and M Lapata (2005). Collective content selection for concept-to-text generation. Proceedings of EMNLP-2005 [1]
  4. ^ R Perera and P Nand (2014). The Role of Linked Data in Content Selection. Proceedings of PRICAI-2014 [2]
  5. ^ J Moore and C Paris (1993). Planning Text for Advisory Dialogues: Capturing Intentional and Rhetorical Information Using. Computational Linguistics 19:651-694 [3]
  6. ^ J Yu, E Reiter, J Hunter, C Mellish (2007). Choosing the content of textual summaries of large time-series data sets. Natural Language Engineering 13:25-49
  7. ^ P Gervás, B Díaz-Agudo, F Peinado, R Hervás (2005) Story plot generation based on CBR. Knowledge-Based Systems 18:235-242