Indexing fails due to immense field

Platform Notice: Data Center Only - This article only applies to Atlassian products on the Data Center platform.

Note that this KB was created for the Data Center version of the product. Data Center KBs for non-Data-Center-specific features may also work for Server versions of the product, however they have not been tested. Support for Server* products ended on February 15th 2024. If you are running a Server product, you can visit the Atlassian Server end of support announcement to review your migration options.

*Except Fisheye and Crucible

Summary

Problem

Both background and foreground reindex requests fail due to a field that has more than 32766 bytes in the field.

The following appears in the atlassian-jira.log

1 2 2019-07-02 16:32:24,247 IssueIndexer:thread-9 WARN 10.x.x.x /secure/admin/jira/IndexReIndex!reindex.jspa [c.a.jira.index.AccumulatingResultBuilder] Document contains at least one immense term in field="<fieldname>" (whose UTF8 encoding is longer than the max length 32766), all of which were skipped. Please correct the analyzer to not produce such terms. The prefix of the first immense term is: '[list of entries in the field]...', original message: bytes can be at most 32766 in length; got 32767 2019-07-02 16:32:24,248 JiraTaskExectionThread-2 WARN 10.x.x.x /secure/admin/jira/IndexReIndex!reindex.jspa [c.a.jira.index.AccumulatingResultBuilder] Indexing failed for Issue - '<#####>'

Diagnosis

Environment

  • This affects Jira 8.0 and above.

  • In Jira 8.0, Lucene was upgraded from 3.3 to 7.3. Instead of silently failing, Lucene now throws an exception.

  • See Lucene upgrade for more information.

Cause

Lucene limits the length of indexed fields to 32766 bytes. Fields that are continually incremented (such as SLA, individual permissions, or custom fields) are more likely to exceed the length limit. The application and the database allow these to increment and do not limit the field length, but Lucene cannot parse immense fields that exceed the byte length, so it fails to index them.

In older versions of Lucene, it would silently fail, but after LUCENE-5472, this will throw an exception.

Solution

Workaround

There is no workaround at this time.

There is bug report: JSWSERVER-20133 - Fix the Lucene immense field indexing failure

Resolution

In order to allow indexing to run, you must either truncate, disable, or delete the problematic field.

  • If the field is managed by a plugin, you can disable that plugin.

  • Limit the input points. For example, if the error indicates a permission field, limit the individuals or groups that are included in the permission field.

  • If it is a custom field that is no longer needed, remove it from active screens, then trim it in the database.

Query the database

Always back up your data before modifying the database. If possible, try your modifications on a test server.

If you would like to query your database on what values are hitting this byte size threshold, you can query the database. Typically, this often happens in the customfield table.

Below is a query tested on PostgreSQL to check the byte size for a given custom field ID:

1 2 3 4 5 6 7 8 9 SELECT pg_column_size(cv.stringvalue) as stringbytesize,pg_column_size(cv.numbervalue),pg_column_size(cv.textvalue) as textbytesize,p.pkey AS "Project", ji.id AS "Issue Key" FROM customfield c JOIN customfieldvalue cv ON c.id = cv.customfield JOIN jiraissue ji ON cv.issue = ji.id JOIN project p ON ji.project = p.id WHERE cv.customfield = examplefield ORDER BY length(stringvalue) desc; -- examplefield should be a numeric value of your customfield_id eg: 10001. stringvalue can be replaced for numbervalue or textvalue.

Updated on April 2, 2025

Still need help?

The Atlassian Community is here for you.