Ranged queries

Main query, which needs to fetch all the documents, can impose a read lock on the whole table and stall the concurrent queries (eg. INSERTs to MyISAM table), waste a lot of memory for result set, etc. To avoid this, Manticore supports so-called ranged queries. With ranged queries, Manticore first fetches min and max document IDs from the table, and then substitutes different ID intervals into main query text and runs the modified query to fetch another chunk of documents. Here’s an example.

Ranged query usage example:

  1. sql_query_range = SELECT MIN(id),MAX(id) FROM documents
  2. sql_range_step = 1000
  3. sql_query = SELECT * FROM documents WHERE id>=$start AND id<=$end

If the table contains document IDs from 1 to, say, 2345, then sql_query would be run three times:

  1. with $start replaced with 1 and $end replaced with 1000;
  2. with $start replaced with 1001 and $end replaced with 2000;
  3. with $start replaced with 2001 and $end replaced with 2345.

Obviously, that’s not much of a difference for 2000-row table, but when it comes to indexing 10-million-row table, ranged queries might be of some help.

sql_query_range

Defines the range query. The query specified in this option must fetch min and max document IDs that will be used as range boundaries. It must return exactly two integer fields, min ID first and max ID second; the field names are ignored. When enabled, sql_query will be required to contain $start and $end macros. Note that the intervals specified by $start..$end will not overlap, so you should not remove document IDs that are exactly equal to $start or $end from your query.

sql_range_step

Defined the range query step. Default is 1024.

sql_ranged_throttle

Can be used to throttle ranged query. By default there is no throttling. Values are expected in sql_ranged_throttle.

Throttling can be useful when indexer imposes too much load on the database server. It causes the indexer to sleep for given amount of milliseconds once per each ranged query step. This sleep is unconditional, and is performed before the fetch query.

  1. sql_ranged_throttle = 1000 # sleep for 1 sec before each query step

Fetching from XML streams

xmlpipe2 source type lets you pass arbitrary full-text and attribute data to Manticore in yet another custom XML format. It also allows to specify the schema (ie. the set of fields and attributes) either in the XML stream itself, or in the source settings.

Declaration of XML stream

xmlpipe_command directive is mandatory and contains the shell command invoked to produce the XML stream which gets indexed. The command can just read a XML file but it can also be a program that generates on-the-fly the XML content.

XML file format

When indexing xmlpipe2 source, indexer runs the given command, opens a pipe to its stdout, and expects well-formed XML stream. Here’s sample stream data:

Example of XML document stream with schema built-in:

  1. <?xml version="1.0" encoding="utf-8"?>
  2. <sphinx:docset>
  3. <sphinx:schema>
  4. <sphinx:field name="subject"/>
  5. <sphinx:field name="content"/>
  6. <sphinx:attr name="published" type="timestamp"/>
  7. <sphinx:attr name="author_id" type="int" bits="16" default="1"/>
  8. </sphinx:schema>
  9. <sphinx:document id="1234">
  10. <content>this is the main content <![CDATA[and this <cdata> entry
  11. must be handled properly by xml parser lib]]></content>
  12. <published>1012325463</published>
  13. <subject>note how field/attr tags can be
  14. in <strong> class="red">randomized</strong> order</subject>
  15. <misc>some undeclared element</misc>
  16. </sphinx:document>
  17. <sphinx:document id="1235">
  18. <subject>another subject</subject>
  19. <content>here comes another document, and i am given to understand,
  20. that in-document field order must not matter, sir</content>
  21. <published>1012325467</published>
  22. </sphinx:document>
  23. <!-- ... even more sphinx:document entries here ... -->
  24. <sphinx:killlist>
  25. <id>1234</id>
  26. <id>4567</id>
  27. </sphinx:killlist>
  28. </sphinx:docset>

Arbitrary fields and attributes are allowed. They also can occur in the stream in arbitrary order within each document; the order is ignored. There is a restriction on maximum field length; fields longer than 2 MB will be truncated to 2 MB (this limit can be changed in the source).

The schema, ie. complete fields and attributes list, must be declared before any document could be parsed. This can be done either in the configuration file using xmlpipe_field and xmlpipe_attr_XXX settings, or right in the stream using <sphinx:schema> element. <sphinx:schema> is optional. It is only allowed to occur as the very first sub-element in <sphinx:docset>. If there is no in-stream schema definition, settings from the configuration file will be used. Otherwise, stream settings take precedence. Note that the document id should be specified as a property id of tag <sphinx:document> (e.g. <sphinx:document id="1235">) and is supposed to be a unique signed positive non-zero 64-bit integer.

Unknown tags (which were not declared neither as fields nor as attributes) will be ignored with a warning. In the example above, \<misc> will be ignored. All embedded tags and their attributes (such as ** in \<subject> in the example above) will be silently ignored.

Support for incoming stream encodings depends on whether iconv is installed on the system. xmlpipe2 is parsed using libexpat parser that understands US-ASCII, ISO-8859-1, UTF-8 and a few UTF-16 variants natively. Manticore configure script will also check for libiconv presence, and utilize it to handle other encodings. libexpat also enforces the requirement to use UTF-8 charset on Manticore side, because the parsed data it returns is always in UTF-8.

XML elements (tags) recognized by xmlpipe2 (and their attributes where applicable) are:

  • sphinx:docset - Mandatory top-level element, denotes and contains xmlpipe2 document set.
  • sphinx:schema - Optional element, must either occur as the very first child of sphinx:docset, or never occur at all. Declares the document schema. Contains field and attribute declarations. If present, overrides per-source settings from the configuration file.
  • sphinx:field - `Optional element, child of sphinx:schema. Declares a full-text field. Known attributes are:
    • “name”, specifies the XML element name that will be treated as a full-text field in the subsequent documents.
    • “attr”, specifies whether to also table this field as a string. Possible value is “string”.
  • sphinx:attr - Optional element, child of sphinx:schema. Declares an attribute. Known attributes are:
    • “name”, specifies the element name that should be treated as an attribute in the subsequent documents.
    • “type”, specifies the attribute type. Possible values are “int”, “bigint”, “timestamp”, “bool”, “float”, “multi” and “json”.
    • “bits”, specifies the bit size for “int” attribute type. Valid values are 1 to 32.
    • “default”, specifies the default value for this attribute that should be used if the attribute’s element is not present in the document.
  • sphinx:document - Mandatory element, must be a child of sphinx:docset. Contains arbitrary other elements with field and attribute values to be indexed, as declared either using sphinx:field and sphinx:attr elements or in the configuration file. The only known attribute is “id” that must contain the unique integer document ID.
  • sphinx:killlist - Optional element, child of sphinx:docset. Contains a number of “id” elements whose contents are document IDs to be put into a kill-list of the table. The kill-list is used in multi-table searches to suppress documents found in other tables of the search

Data definition in source configuration

If the XML doesn’t define a schema, the data types of tables elements must be defined in the source configuration.

  • xmlpipe_field - declares a text field.
  • xmlpipe_field_string - declares a text field/string attribute. The column will be both indexed as a text field but also stored as a string attribute.
  • xmlpipe_attr_uint - declares an integer attribute
  • xmlpipe_attr_timestamp - declares a timestamp attribute
  • xmlpipe_attr_bool - declares a boolean attribute
  • xmlpipe_attr_float - declares a float attribute
  • xmlpipe_attr_bigint - declares a big integer attribute
  • xmlpipe_attr_multi - declares a multi-value attribute with integers
  • xmlpipe_attr_multi_64 - declares a multi-value attribute with 64bit integers
  • xmlpipe_attr_string - declares a string attribute
  • xmlpipe_attr_json - declares a JSON attribute

Specific XML source settings

If xmlpipe_fixup_utf8 is set it will enable Manticore-side UTF-8 validation and filtering to prevent XML parser from choking on non-UTF-8 documents. By default this option is disabled.

Under certain occasions it might be hard or even impossible to guarantee that the incoming XMLpipe2 document bodies are in perfectly valid and conforming UTF-8 encoding. For instance, documents with national single-byte encodings could sneak into the stream. libexpat XML parser is fragile, meaning that it will stop processing in such cases. UTF8 fixup feature lets you avoid that. When fixup is enabled, Manticore will preprocess the incoming stream before passing it to the XML parser and replace invalid UTF-8 sequences with spaces.

  1. xmlpipe_fixup_utf8 = 1

Example of XML source without schema in configuration:

  1. source xml_test_1
  2. {
  3. type = xmlpipe2
  4. xmlpipe_command = cat /tmp/products_today.xml
  5. }

Example of XML source with schema in configuration:

  1. source xml_test_2
  2. {
  3. type = xmlpipe2
  4. xmlpipe_command = cat /tmp/products_today.xml
  5. xmlpipe_field = subject
  6. xmlpipe_field = content
  7. xmlpipe_attr_timestamp = published
  8. xmlpipe_attr_uint = author_id:16
  9. }

Fetching from TSV,CSV

This is the simplest way to pass data to the indexer. It was created due to xmlpipe2 limitations. Namely, indexer must map each attribute and field tag in XML file to corresponding schema element. This mapping requires some time. And time increases with increasing the number of fields and attributes in schema. There is no such issue in tsvpipe because each field and attribute is a particular column in TSV file. So, in some cases tsvpipe could work slightly faster than xmlpipe2.

File format

The first column in TSV/CSV file must be a document ID. The rest ones must mirror the declaration of fields and attributes in schema definition. Note that you don’t need to declare the document ID in the schema, since it’s always considered to be present, should be in the 1st column and needs to be a unique signed positive non-zero 64-bit integer.

The difference between tsvpipe and csvpipe is delimiter and quoting rules. tsvpipe has tab character as hardcoded delimiter and has no quoting rules. csvpipe has option csvpipe_delimiter for delimiter with default value , and also has quoting rules, such as:

  • any field may be quoted
  • fields containing a line-break, double-quote or commas should be quoted
  • a double quote character in a field must be represented by two double quote characters

Declaration of TSV stream

tsvpipe_command directive is mandatory and contains the shell command invoked to produce the TSV stream which gets indexed. The command can just read a TSV file but it can also be a program that generates on-the-fly the tab delimited content.

TSV indexed columns

The following directives can be used to declare the types of the indexed columns:

  • tsvpipe_field - declares a text field.
  • tsvpipe_field_string - declares a text field/string attribute. The column will be both indexed as a text field but also stored as a string attribute.
  • tsvpipe_attr_uint - declares an integer attribute
  • tsvpipe_attr_timestamp - declares a timestamp attribute
  • tsvpipe_attr_bool - declares a boolean attribute
  • tsvpipe_attr_float - declares a float attribute
  • tsvpipe_attr_bigint - declares a big integer attribute
  • tsvpipe_attr_multi - declares a multi-value attribute with integers
  • tsvpipe_attr_multi_64 - declares a multi-value attribute with 64bit integers
  • tsvpipe_attr_string - declares a string attribute
  • tsvpipe_attr_json - declares a JSON attribute

Example of a source using a TSV file

  1. source tsv_test
  2. {
  3. type = tsvpipe
  4. tsvpipe_command = cat /tmp/rock_bands.tsv
  5. tsvpipe_field = name
  6. tsvpipe_attr_multi = genre_tags
  7. }
  1. 1 Led Zeppelin 35,23,16
  2. 2 Deep Purple 35,92
  3. 3 Frank Zappa 35,23,16,92,33,24

Declaration of CSV stream

csvpipe_command directive is mandatory and contains the shell command invoked to produce the TSV stream which gets indexed. The command can just read a CSV file but it can also be a program that generates on-the-fly the comma delimited content.

CSV indexed columns

The following directives can be used to declare the types of the indexed columns:

  • csvpipe_field - declares a text field.
  • csvpipe_field_string - declares a text field/string attribute. The column will be both indexed as a text field but also stored as a string attribute.
  • csvpipe_attr_uint - declares an integer attribute
  • csvpipe_attr_timestamp - declares a timestamp attribute
  • csvpipe_attr_bool - declares a boolean attribute
  • csvpipe_attr_float - declares a float attribute
  • csvpipe_attr_bigint - declares a big integer attribute
  • csvpipe_attr_multi - declares a multi-value attribute with integers
  • csvpipe_attr_multi_64 - declares a multi-value attribute with 64bit integers
  • csvpipe_attr_string - declares a string attribute
  • csvpipe_attr_json - declares a JSON attribute

Example of a source using a CSV file

  1. source csv_test
  2. {
  3. type = csvpipe
  4. csvpipe_command = cat /tmp/rock_bands.csv
  5. csvpipe_field = name
  6. csvpipe_attr_multi = genre_tags
  7. }
  1. 1,"Led Zeppelin","35,23,16"
  2. 2,"Deep Purple","35,92"
  3. 3,"Frank Zappa","35,23,16,92,33,24"