this point in the test since the collections available are changing due
to deletions and we might try to communicate with a collection
that was (correctly) deleted.
(cherry picked from commit a7435d9e6e2b6eac9b855157c969fa97cc447b35)
remove flakey dependency on sleep (to wait for cache expiration) w/explicit instruction to ignore cache
sleep is not garunteed to sleep the exact amount (based on limitations of the system clock)
(cherry picked from commit 24b94b8dac4a8b4aea80294390d92705d164b546)
If the field is non-stored, non-indexed and docvalue enabled numeric field
then inplace update can be done. previously, lucene didn't support
docvalue update for field that is not yet present in indexWriter but
LUCENE-8316 added support for this.
This adds support to update field which satisfies inplace conditions
but which doesn't yet exist in any docs
Remove buggy 'port roulette' code that can easily fail if OS gives the selected port to a different process just before creating the server
Use jetty's built in support for listining on an OS selected port instead
Also increase timeouts to better account for slow/heavily loaded (ie:jenkins) VMs where SolrCore reloading may take longer then 10 seconds
(cherry picked from commit 19c78ddf98b1cef86f7a1c6d124811af8726b41d)
* tighten assertions related to type of watcher that should be removed
* use waitForActiveCollection before deleting collections to work around SOLR-13616 and/or SOLR-13627
(cherry picked from commit 7ddba3b7123f11d387ff395516cf46c607ac21ee)
- ensure all collections/replicas are active
- tighten assertions around expected replica locations
- eliminate some redundent code
These changes should also help ensure we don't get (more) spurious failures due to SOLR-13616
(cherry picked from commit 8a277cab7d15c03ff59577efceb6a0cb281d095a)
Payload func with undefined used to throw NPE. In SOLR-11610, this
was fixed to return proper error but there are no tests to verify
changed the behavior.
This add simple test to verify error code and error message
group.query after execution forms QueryCommandResult. In case of
group.main=true or group.format=simple, QueryCommandResult was not
consumed in EndResultTransformer. Also, MainEndResultTransformer assumed
that always group.field would be specified. When group.field not specified
it failed with AIOOBE. After adding suppport for QueryCommandResult in
EndResultTransformers and handling AIOOBE, group.query started giving results
Working on tests exposed few other issues. Results differed b/w standalone
& distributed mode.
* One of the reason is that TopGroupShardResponseProcessor doesn't consider correct
limit and offset when group format is simple. In case of simple, start and rows should be used
as limit and offset instead of group.limit and group.offset.
* Secondly, In distributed second phase grouping, computing docsToCollect didn't consider
group response format. This issue is again similar to above issue
* offset(group.offset or start) not being considered during TopDocs#merge caused
different results. The fix was to use to offset in merge process
* group.offset doesn't support negative values but there is no checks on the value.
In case of negative values AIOOBE. Now, checks are added for negative values and
returns proper error message(this change is for both standalone and distrbuted).
Validation is done only in case of group.format=grouped as that is only case when
group.offset is consumed.
Fixing above issues resolved the differences b/w standalone and distributed mode.
* When ramPerThreadHardLimitMB is not specified, then Lucene's
default value 1945 is used. The specified value should be
greater than 0 and less than 2048MB