Let’s say the staging directory is
/mnt/photo/staging/NewZealand
Assume that the database is in files.db
python3 manage_dupes.py --source /mnt/photo/staging/NewZealand/ --dest /mnt/data/new_photo_dupes/NewZealand --db-path files.db --action move --verbose 3 --literal-patterns --skip-pattern .DS_Store --skip-pattern .picasa.ini -skip-pattern Thumbs.db --dry-run
Using Digikam:
Location
Set Content→Captions
Origin→City, State, Country, Address
Keywords → E.g Event/Name Trips/ …
Properties→Priority
python3 ~/monorepo/scripts/index_media.py --path /mnt/photo --check-existing size --check-existing modified_date --check-existing fullpath --volume PHOTO --db-path /home/ubuntu/data/media-index/files.db --literal-patterns --skip-pattern "/.filerun" --skip-pattern "/.DS_Store" --start-dir 2018 --start-dir 2019 --start-dir 2020 --verbose 2
Index the directories 2019 and 2020 under /mnt/photo
Catch duplicates based on modified_date, fullpath and size
python3 remove_dupes.py --dest /mnt/photo/photo_dupes/ --db-path ~/data/media-index/files-2026-01-25.db --base-dir /mnt/photo --dry-run --verbose 3
Moves the dupes found into /mnt/photo/photo_dupes. Base Dir is removed from full path to create the relative path for the photo_dupes folder
rsync -avh --progress --partial --inplace /Volumes/External/Export /Volumes/photo/PhotosExport
where
/Volumes/External/Export : the export folder on the Mac
/Volumes/photo/PhotosExport: The smb mounted folder from NAS