Description
This issue adds a new TolerantUpdateProcessorFactory making it possible to configure solr updates so that they are "tolerant" of individual errors in an update request...
<processor class="solr.TolerantUpdateProcessorFactory"> <int name="maxErrors">10</int> </processor>
When a chain with this processor is used, but maxErrors isn't exceeded, here's what the response looks like...
$ curl 'http://localhost:8983/solr/techproducts/update?update.chain=tolerant-chain&wt=json&indent=true&maxErrors=-1' -H "Content-Type: application/json" --data-binary '{"add" : { "doc":{"id":"1","foo_i":"bogus"}}, "delete": {"query":"malformed:["}}' { "responseHeader":{ "errors":[{ "type":"ADD", "id":"1", "message":"ERROR: [doc=1] Error adding field 'foo_i'='bogus' msg=For input string: \"bogus\""}, { "type":"DELQ", "id":"malformed:[", "message":"org.apache.solr.search.SyntaxError: Cannot parse 'malformed:[': Encountered \"<EOF>\" at line 1, column 11.\nWas expecting one of:\n <RANGE_QUOTED> ...\n <RANGE_GOOP> ...\n "}], "maxErrors":-1, "status":0, "QTime":1}}
Note in the above example that:
- maxErrors can be overridden on a per-request basis
- an effective maxErrors==-1 (either from config, or request param) means "unlimited" (under the covers it's using Integer.MAX_VALUE)
If/When maxErrors is reached for a request, then the first exception that the processor caught is propagated back to the user, and metadata is set on that exception with all of the same details about all the tolerated errors.
This next example is the same as the previous except that instead of maxErrors=-1 the request param is now maxErrors=1...
$ curl 'http://localhost:8983/solr/techproducts/update?update.chain=tolerant-chain&wt=json&indent=true&maxErrors=1' -H "Content-Type: application/json" --data-binary '{"add" : { "doc":{"id":"1","foo_i":"bogus"}}, "delete": {"query":"malformed:["}}' { "responseHeader":{ "errors":[{ "type":"ADD", "id":"1", "message":"ERROR: [doc=1] Error adding field 'foo_i'='bogus' msg=For input string: \"bogus\""}, { "type":"DELQ", "id":"malformed:[", "message":"org.apache.solr.search.SyntaxError: Cannot parse 'malformed:[': Encountered \"<EOF>\" at line 1, column 11.\nWas expecting one of:\n <RANGE_QUOTED> ...\n <RANGE_GOOP> ...\n "}], "maxErrors":1, "status":400, "QTime":1}, "error":{ "metadata":[ "org.apache.solr.common.ToleratedUpdateError--ADD:1","ERROR: [doc=1] Error adding field 'foo_i'='bogus' msg=For input string: \"bogus\"", "org.apache.solr.common.ToleratedUpdateError--DELQ:malformed:[","org.apache.solr.search.SyntaxError: Cannot parse 'malformed:[': Encountered \"<EOF>\" at line 1, column 11.\nWas expecting one of:\n <RANGE_QUOTED> ...\n <RANGE_GOOP> ...\n ", "error-class","org.apache.solr.common.SolrException", "root-error-class","java.lang.NumberFormatException"], "msg":"ERROR: [doc=1] Error adding field 'foo_i'='bogus' msg=For input string: \"bogus\"", "code":400}}
...the added exception metadata ensures that even in client code like the various SolrJ SolrClient implementations, which throw a (client side) exception on non-200 responses, the end user can access info on all the tolerated errors that were ignored before the maxErrors threshold was reached.
Has anyone run into the problem of handling bad documents / failures mid batch. Ie:
<add>
<doc>
<field name="id">1</field>
</doc>
<doc>
<field name="id">2</field>
<field name="myDateField">I_AM_A_BAD_DATE</field>
</doc>
<doc>
<field name="id">3</field>
</doc>
</add>
Right now solr adds the first doc and then aborts. It would seem like it should either fail the entire batch or log a message/return a code and then continue on to add doc 3. Option 1 would seem to be much harder to accomplish and possibly require more memory while Option 2 would require more information to come back from the API. I'm about to dig into this but I thought I'd ask to see if anyone had any suggestions, thoughts or comments.
Attachments
Attachments
Issue Links
- depends upon
-
SOLR-8633 DistributedUpdateProcess processCommit/deleteByQuery call finish on DUP and SolrCmdDistributor, which violates the lifecycle and can cause bugs.
- Resolved
- is blocked by
-
SOLR-8738 invalid DBQ initially sent to a non-leader node will report success
- Resolved
- is duplicated by
-
SOLR-7914 Improve bulk doc update
- Closed
- is part of
-
SOLR-3382 Finegrained error propagation (focus on multi-document updates)
- Open
- is related to
-
SOLR-3178 Versioning - optimistic locking
- Open
-
SOLR-8881 test & document (and improve as possible) behavior of TolerantUpdateProcessor while shard splitting is in progress
- Open
-
SOLR-1113 Error reports from ExtractingRequestHandler and Co do not indicate name of rejected documents
- Closed
-
SOLR-8872 ChaosMonkey depends on AbstractFullDistribZkTestBase, can't be used with MiniSolrCloudCluster
- Open
- links to