You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Following error has been observed when a single value for a field is very large:
Caused by: org.elasticsearch.hadoop.rest.EsHadoopInvalidRequest: Found unrecoverable error [141.211.168.139:9200] returned Bad Request(400) - Document contains at least one immense term in field="dc_subject.keyword" (whose UTF8 encoding is longer than the max length 32766), all of which were skipped. Please correct the analyzer to not produce such terms. The prefix of the first immense term is: '[72, 111, 103, 103, 44, 32, 74, 97, 109, 101, 115, 32, 82, 46, 59, 32, 85, 110, 105, 118, 101, 114, 115, 105, 116, 121, 32, 69, 118, 101]...', original message: bytes can be at most 32766 in length; got 38743; Bailing out.
Sounds like removing the long value in the XML fixed this indexing error, but should gracefully handle even extremely long fields (e.g. full-text transcription for a book).
Possible approaches:
add flag to xml2kvp that would set truncate length for a field
if assuming that index mapping is largely about analysis, and ES is not designed for storing data in that way, not a bad option
hard limit: skip the flag, just truncate below the 32766 length
The text was updated successfully, but these errors were encountered:
Following error has been observed when a single value for a field is very large:
Sounds like removing the long value in the XML fixed this indexing error, but should gracefully handle even extremely long fields (e.g. full-text transcription for a book).
Possible approaches:
32766
lengthThe text was updated successfully, but these errors were encountered: