Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

sycamore_importer: manage deleted and updated documents #156

Open
eric-anderson opened this issue Nov 9, 2023 · 0 comments
Open

sycamore_importer: manage deleted and updated documents #156

eric-anderson opened this issue Nov 9, 2023 · 0 comments

Comments

@eric-anderson
Copy link
Collaborator

Right now the importer has no way to notice that a crawled file has been deleted and remove it from the index. Similarly, if a file is updated, the crawlers will detect that, re-download the file and the importer will then import it again without removing the old file resulting in duplicate entries in the index.

It's unclear what the right way to handle this is, but a likely choice is that the files stored in the crawl_data compose volume should match the ones stored in opensearch and there should be no duplicates.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant