so many duplicates, how to break it up into smaller chunks or different kinds of duplication?

I have a lot of duplicates to work through.

I found the zoplicate plugin that works on Zotero 7. Even as manual review goes, it offers the benefit of being able to make "not a duplicate" which is a big missing piece in the native functionality because otherwise you end up looking at the some items over and over again. But I don't want to point it at my collection for bulk merge because it seems destructive and difficult to tell either before or after what it would do or did, given how many potential duplicates I have.

Is there any way to sort, filter duplicates, or break them off in chunks?

In my collection, about 50% of the duplicates identified should be easy for the computer to have high confidence in because they are:

- Exact duplicates - there is no difference at all between the items except the time stamp

Another 30% are items that I'd like to double check if possible but would be happy to automatically merge if it were the only option:

- Difference in URL (got the same item twice from 2 different places)

And items which are higher probability of false positives
- difference in page numbers
- difference in volume, issue numbers
- Other clashes in meta data

Failing the above, I would take any method to split off 50 items at a time and work on them in a chunk. Right now my task just feels hopeless.

Sign In or Register to comment.