* prepare to add $delete-expunge operation that will create a spring batch job
* Add operation
* Wire up jpa provider. Begin with failing test.
* Copy/paste bulk import job as a starting point.
FIXME with proposed design
* delete expunge job parameter validation with test
* implemented reader
stubbed processor, writer
* wip for master merge
* started implementing reader
* started implementing reader
* working with stubs
* happy path batch delete expunge is done
* Provider done but test not passing. Guessing batch infrastructure not running in that test.
* IT test works now
* add reader test
* Converted delete _expunge=true to use new batch job
* DeleteExpungeDaoTest passes
* Fix test
* Change batch size to integer
* rename search count to batch size
* Make delete expunge partition aware
* updated docs
* pre-review cleanup
* change log
* add partition id to SystemRequestDetails
* Make RequestPartitionId serializable
* Change delete expunge provider to use partition id instead of tenant name
* fix tests
* test pointcut gets called
* assert on pointcut calls
* Add resource type to STORAGE_PARTITION_SELECTED pointcut
* bump hapi-fhir version
move expunge provider parameters from JpaConstants to ProviderConstants
* bump hapi-fhir version
* copyrights
* restore deleteexpungeservice for mdm
* restore deleteexpungeservice for mdm
* fix test
* public constants
* convert instant to date
* Moved expunge constants to ProviderConstants
* final review
* disabling InMemoryResourceMatcherR5Test.testNowNextMinute() to see if I can get a clean test run
* fix tests
* fix tests
* fix tests
* fix tests
* review feedback
* review feedback
* review feedback
* review feedback
* review feedback
* review feedback
* improve logging
* bump version
* version bump
* recovering from failed merge
* unzip RequestListJson per Gary's suggestion. I didn't want to do it at first, but as usual Gary was right.
* fix serialization
* Work on changes
* Work on perf
* Work on testing
* Work on perf
* Work on perf
* Work on fix
* Work on perf
* Ongoing work
* Add changelog
* Additional docs
* Test fixes
* Address review comments
* Test fix
* Address DOS issue in History operation
* Work on history caching
* Add docs
* Add changelog
* Resolve test failures
* Test fixes
* Test fix
* Test fixes
* Start work on bul;k import
* Work on bulk import
* Have batch working
* Working
* Working
* More work
* More work on bulk export
* Address fixmes
* License header updates
* Test fixes
* License header updates
* Test fix
* Test fix
* Version bumps
* Work on config
* Test cleanup
* One more version bump
* Version bump
* CLeanup
* A few additions
* Test fixes
* Test fix
* Test fix
* Migration fix
* Test fix
* Test fix
* Start work
* Work on testing
* Version bump
* Test fix
* Compile fix
* Test Fixes
* Test fix
* Test fix
* Compile fix
* Tst fix
* Test fix
* Test fix
* Test fix
* Added index to improve the performance
* Added index for FORCED_ID
* Added comments for the index of while 'missing' is supported.
* Updated comments
* Moved the migration script to V5_4_0
* Automatic reference versions
* Auto reference versions
* resolve fixmes
* Add changelog
* Test fixes
* Test fixes
* Resolve fixmes
* Work on versioning
* Add support for _includes
* Update headers
* Change visibility
* One more tweak
* One more API tweak
* Fixes
* Account for review comments
* Fix errorprone reported issue
* Test fix
* Test fixes
* Test fixes
* switched to adding annotations. Just did a test with Token, but the sql looks clean.
* switched to adding annotations. Just did a test with Token, but the sql looks clean.
* added the other six index links
* trying different annotations
* Start testing
* Update resources on package install
* Add changelog
* Join rework
* CLean up SQL builder
* Add docs
* SP rework
* Join work
* Work on params
* Work on refactor
* Work on chains
* Work on joins
* Rework queries
* Work on queries
* Many more tests passing
* Refs test
* Work on sorting
* Work on tests
* More joins work
* Work on tests
* Work on queries
* Tests passing
* More test fixes
* Test fixes
* Work on SQL
* Tests passing
* Add some tests
* Add some tests
* License headers
* Use entity manager to get datasourcd
* One more fix
* Model cleanup
* Ongoing work
* Fixes
* Fixes
* Work on joins
* Ongoing fixes
* Merge conflict
* Cleanup
* clean up unused fields
* Work on join
* COmpile fix
* Rework querying
* Update hapi-fhir-jpaserver-base/src/main/java/ca/uhn/fhir/jpa/config/HibernateDialectProvider.java
Co-authored-by: Ken Stevens <khstevens@gmail.com>
* Address review comments
* Resolve fixmes
* Test fix
* Test fixes
* Test fix
Co-authored-by: Ken Stevens <khstevens@gmail.com>
* Improve synchronous searching by providing offset & limit support
Add support for offset querying which leverages paging at the query level
Add configuration for search default page size and search maximum page size
If using offset, always use synchronous searching to avoid extra database insert/update etc.
When using offset, only calculate count if it's wanted
Validate params closer to provider and handle size if search returned "all" (or last)
* Review and test fixes
Comment _offset as nonstandard parameter
Make synchronous search always count the total value (for now)
Fix issue with FulltextSearchSvcImpl mutating param map
Dirty fix for BaseJpaTest (not sure how to fix the including resources issue)
* Remove temporary count querying fix for synchronous loads
* Fix offsetting of everything operations, do not drop zero offset
* Fix jpa test default and maximum page size, add some fixmes to tests before resolved
* Ignore one failing test, fix others
* Fix Dereferenced variable may be null
* Fix everything paging in R4 by adding DISTINCT if synchronous load is used
Also fix assertion of size when hitting fetchSizeDefaultMaximum
* Add documentation about offset annotation and paging
Co-authored-by: James Agnew <jamesagnew@gmail.com>