Duplicacy prune Why I can’t Hi, i just performed a backup as a test and later i did duplicacy prune -exclusive -r 1 The revision is removed but in the chunks dir on the storage i still see some empty chunk dirs. I’m using the following options: -exclusive -keep 0:90 -keep 7:30 -keep 1:3 -ignore clonezilla -ignore scyllams_part The results are that nothing gets pruned in the storage, and I suspect I’m using the “-ignore” option incorrectly. The user guide for the old GUI version can be found here. Yes, you can remove the subdirectory under the snapshots directory, but this will leave many unreferenced chunks on the storage. Hi, looking for a bit of help please. All seems to work fine except that the cloud storage grows in size despite pruning. Please describe what actually happens (the wrong behaviour): Output of Found redundant chunk floor(# of chunks Delete old snapshots from snapshots sub folder. There is safety I have a local and remote storage. Any and all help is appreciated as this is the first time I’m trying to interact with the CLI to do anything. Explanation given is that prune deletes files in two steps: first renaming chunks to fossils, then actually removing only if there For this my goal is to run duplicacy every hour, have it email me any errors, and then prune (thin) the backed up data. On Windows and macOS, Duplicacy Web Edition is provided as installers. All duplicacy command (except the copy command) only work with one storage at a time. But for some machines I feel like it would be more suitable to simply keep, for instance, the last 3 snapshots, rather than consider when those I have been using duplicacy for a week now It is absolutely amazing. This is because for chunks to be deleted, the first prune collects prospective chunks to be deleted, then waits for at least one backup to be completed for all repositories to be sure they’re no longer needed (any that are, Now I want to completely delete all backup revisions of a specific backup id. This appears to have fixed the issue and the entire backup job is now running in 2-4 hours (that includes creating/copying images of two server volumes, backing up the server target volume to the local Duplicacy storage pool, pruning both the local and cloud storage pools, copying the local storage pool to the cloud with snapshots from four targets). my checks kept failing for months and I never got any emails, and the prunes kept I have roughly 350 revisions. Reference: Running prune command from /cache/localhost/all Options: [-log prune -storage hidden -a -threads 10 -keep 0:7] 2021-11-17 01:01:33. Is this intended? If yes, why? This Empty folders after prune partially answer my question. Click here for a list of related forum topics. Quick overview NAME: duplicacy prune - Prune revisions by number, I would like to set up Duplicacy on 8 Windows computers (small company) to back them up completely to Google G Suite Drive (business plan). I'm curious to know For those of us with a finite amount of space and a closed wallet (almost everyone), pruning makes complete sense - not least because having too many snapshots, NAME: duplicacy prune - Prune revisions by number, tag, or retention policy USAGE: duplicacy prune [command options] OPTIONS: -id <snapshot id> delete revisions with the specified snapshot ID instead of the default one -all, -a match against all snapshot IDs -r <revision> [+] delete the specified revisions -t <tag> [+] delete revisions with the specified tags Check if the chunk was not deleted by prune. So: duplicacy list Snapshot hps_pool2g revision 1 created at 2017-10-03 16:56 -hash Snapshot hps_pool2g revision 2 created at 2017-10-06 14:08. 3: 63: 10 December 2024 Weekly / monthly scheduled tasks in duplicacy_web. Thanks! I take that as a “yes” regarding the last question . Your usecase as described is pretty straightforward: init repo, do backup, run check and/or restore. I have read the page on wiki and pdf with detailed explanation of how pruning works. -keep 30:14 applies to backup snapshots older than 14 days and keeps one version every 30 days. If you wanted to actually remove the unreferenced chunks, then remove the -dry-run option. On local storage the backup size is ca. The add command is similar to the init command, except that the first argument is a storage name used to distinguish different storages: Running the web-based GUI (Duplicacy Web Edition) to restore, check, copy, or prune existing backups You can use the personal GUI license to back up any files if you are a student, or a staff/faculty member at a school or university. Reading through the prune help pages and various other posts, I think my settings are correct: -keep 0:14 -keep 7:7 -keep 1:1 Which should translate to: -keep 0:14 # Is snapshot older that 14 days? Yes keep 0 revision List of storage names to prune for duplicacy prune operations* check: List of storage names to check for duplicacy check operations* Note that * denotes that this section is mandatory and MUST be specified in the configuration file. Quick overview NAME: Then click on the prune options there and replace the -keep arguments with what you actually wanted. Prune ran and completed jus I have been very impressed with Duplicacy. The main beef is Pruning. Snapshot Pruning. In the GUI I set non-default settings for Rate Limit, Threads, Volume Shadow Copy Schedule, and I turned on Prune Snapshots. exe prune -id laptop -keep 365:365 -keep 30:90 -keep 7:10 -keep 1:2 -threads 64 Please describe what you expect to happen (but doesn’t): Delete old snapshot according to the criteria. It seems it works as designed. This is the full-featured web-based GUI frontend supporting all major operations (backup, restore, check, copy, and prune) as well as other features such as scheduling and email notifications. So far everything works great but I am confused with the prune settings. I am totally new to Duplicacy so I am sorry if I am asking an idiotic question that is obvious to all. It also integrates a dashboard to provide a graphical view of the status of storages and backups. Let’s go through this piece by piece. Of course it fails. And for snapshots newer than 7 days it keeps every snapshot? So I have daily snapshots for 1 week, and 4 Duplicacy est un puissant outil de sauvegarde dans le cloud qui offre des capacités de sauvegarde et de déduplication de données efficaces et sécurisées. link or the plus sign at the top right corner. robert. Remember that no other backup can be running when you run this command (due to the use of -exclusive). The important thing is that there are a number of backup ids which i stopped using 2-3 years ago and waited for them to gradually get pruned, You can delete orphaned snapshots manually (from subfolder “snapshots”) on OneDrive, then clear local duplicacy cache, and run prune -exhaustive to cleanup orphaned chunks. Bluebeep Jan 7 1:04PM 2018 GUI. I downloaded and installed the Dublicacy Web Edition on a Windows x64 machine and set up a couple of backup jobs. I know there have been other threads about understanding the Prune command, but after searching through them they’re still not resolving my confusion or, rather, not in a way I understand so I would hugely appreciate some direct help! This is the setup: I have With duplicacy in current state of affairs I don’t use prune. It has now run for a day and, as expected, Hi, I’m just starting with Duplicacy and would like to hear the best practice for when I need to have a Check & Prune job in my time schedule. Understanding Snapshots and Pruning. (Duplicacy Web Edition) to restore, check, copy, or prune existing backups So you can set up the backups using the CLI, and still have the pretty restore screen for when you So I’ve created a prune schedule on it with the flag -keep 0:30 The 30 days have passed however and the email alerts I’m getting from Duplicacy are saying: INFO RETENTION_POLICY Keep no snapshots older than 30 days INFO SNAPSHOT_NONE No snapshot to delete So the policy is being applied but nothing is being done? Running prune Add prune job with the desired schedule and ignore retention settings configurable in the dialog. Il est conçu pour simplifier le processus de sauvegarde tout en optimisant l'espace de stockage en éliminant les données redondantes. Please describe what actually happens (the wrong behaviour): It writes a long list of Deleting snapshot laptop I’m backing up several machines to my duplicacy repo but some of them somewhat sporadically. pdf at master · gilbertchen/duplicacy · GitHub but the gist of it is that duplicacy is fully concurrent, it has to support weird scenarios where five machines run backup to a specific storage and seven other machines try to prune it at the same time. Which has worked great. duplicacy/logs. I think the appropriate command would be duplicacy prune -all -exhaustive -dry-run. 498 INFO RETENTION_POLICY Keep no I installed the Duplicacy docker to backup my unraid system to an old synology nas at my mom´s house. but over the last few days i've had to split out one of the sources to 3 seperate ones. Yes. I have only seen it with -keep 1:1. Feature. backblazeb2. If I never prune, all is fine, backups A place for Duplicacy users and developers. Prune options are not set by default. 1 seems to have a less verbose Everything under . Duplicacy’s -keep command controls what to keep – e. Check: Daily, early morning before the first backup. Wasabi’s 90-day minimum for stored data means there is no financial incentive to reduce utilization through early pruning of snapshots. But I am still unable to quite fully understand how it works and how is it able to work lock-free. 0 prune -exhaustive’ to remove chunks associated with the now-deleted snapshot, presuming I can also find the preferences that the duplicacy command wants. Ok. Data is not lost — but checks will start failing on the ghost snapshots that were supposed to be deleted. Duplicacy comes with a newly designed web-based GUI that is not only artistically appealing but also functionally powerful. The prune command is sent via WebUI: “-id S_Bilder -r 1-1000 -exclusive” (1) duplicacy-web does not send an email on failure for a check task (2) if a schedule has a check task and a prune task, and if the check task fails (e. The reasons — it leaves datastore in bad state if interrupted. They are combinable – from oldest to youngest: duplicacy prune -keep 0:360 -keep 30:180 -keep 7:30 $ duplicacy prune # Chunks from deleted backups will be removed if deletion criteria are met To back up to multiple storages, use the add command to add a new storage. There are 4 types of jobs: backup, check, copy and prune. That’s it, simple. If the Prune snapshot after first backup checkbox is checked, Duplicacy will run the pruning operation after the first backup of the day or the manual backup. I am using the Web GUI in a Docker environment. muench 19 February 2019 08:50 #3. 1. I have this command running every To summarize: if you did not violate duplicacy assumptions: did not touch the same snapshot ID from two different places, did not use -exclusive flag, did not interrupt prune, there shall be no way to silently lose data. You can select the second one by New Duplicacy user here, running the Web-UI version in a Docker environment from my Synology NAS. RUN_JOB_IMMEDIATELY: Set to yes to run duplicacy backup and/or duplicacy prune command at container startup. I currently schedule a backup, check and prune daily, but having issues where it seems old revisions aren’t deleting. I’m trying to figure out Duplicacy’s versioning/retention settings (and their limitations). Quick overview NAME: duplicacy prune - Prune snapshots by revision, tag, or retention policy USAGE: duplicacy prune [command options] OPTIONS: -id <snapshot id> delete snapshots with the specified id instead of the default one -all, -a match against all snapshot IDs -r <revision> [+] delete -all doesn’t mean to prune all the storages in the preferences file. The storage list If running prune a second time didn’t work, you may have to run it with the -exclusive flag, making sure no backups are running while it does. Prune: Weekly or bi-weekly, based on your data change rate. Please note that if there are multiple Duplicacy instances running on different [root@mail ~]# . However, there is one thing I think should be handled differently: the behaviour of prune -keep . I tried again, this time Starting each duplicacy prune -keep 0:360 -keep 30:180 -keep 7:30 -keep 1:7 1 Like. You can add -exclusive flag to do it quicker but ensure there is no other duplicacy process interacting with the datastore at the same time. What does Pruning do exactly? The Keep 1 snapshot feature makes sense however duplicacy prune - Prune revisions by number, tag, or retention policy. Replace delete all -keep arguments and replace them with your own -keep 0:2. I find it unacceptable, and until it’s fixed I’m not using prune. Without prune the datastore is add-only. This is because for chunks to be deleted, the first prune collects prospective chunks to be deleted, then waits for at least one backup to be completed for all repositories to be sure they’re no longer needed (any that are, . Any guidance on how to I continue to test Duplicacy via its GUI on a second, Win 7, PC. com 2021-11-17 01:01:35. Say I run three backups a day and once every week I prune old backups with the following: duplicacy prune -keep 1:2 -all Duplicacy will keep one revision per day for revisions older than 2 days, fine. I run a backup and check every night. You can use this list · gilbertchen/duplicacy Wiki · GitHub to find out which revisions contain specific file. I have the following options enabled on a prune command set to run after every scheduled backup: -d | -keep 7:7 -keep 0:30 -a -exhaustive -threads 10 I interpret this as follows: Keep one revision every 7 days after seven Please describe what you are doing to trigger the bug: I run duplicacy. Most If you’ve already deleted the unused snapshots under the snapshots directory in your storage, then you can use a prune to identify unreferenced chunks. Loses all changes between oldest and youngest revisions. After a job has been added, you can set the There are 3 types of Duplicacy licenses: Running the CLI version to restore, check, copy, or prune existing backups on any computer; Running the CLI version to back up any files on a computer where a GUI license is already installed; Running the web-based GUI (Duplicacy Web Edition) to restore, check, copy, or prune existing backups After that, run duplicacy prune -exclusive -exhaustive to clean up the storage. Moreover, since most cloud storage services do not provide a locking service, the best effort is to use some basic file Will this work: duplicacy prune -keep 0:20. Duplicacy offre des fonctionnalités telles que la gestion des versions, le SYNOPSIS: duplicacy check - Check the integrity of snapshots USAGE: duplicacy check [command options] OPTIONS: -all, -a check snapshots with any id -id <snapshot id> check snapshots with the specified id rather than the default one -r <revision> [+] the revision number of the snapshot -t <tag> check snapshots with the specified tag -fossils search fossils if a chunk If running prune a second time didn’t work, you may have to run it with the -exclusive flag, making sure no backups are running while it does. 1: 270: 10 December 2024 Not able to renew license. duplicacy-util duplicacy prune -keep 0:360 -keep 30:180 -keep 7:30 -keep 1:7 Second question. This sounds like a good idea if I wanted to prune stale or deprecated repos manually. There’s always something failing. As a test I configured the jobs to backup a user’s folder tree and one critical software specific folder tree. Today I got a warning that I am exceeding my limits. Although there is supposedly unlimited space available on Google Drive, I would I've been running duplicacy for a while now. Depends on your target storage performance (one drive — slow, Amazon s3 — fast), amount of data turnover Under the hood, it will create a directory named . “I’d definitely not tinker with the /chunks directory if I were you - if you still have any backup data in A prune operation will completely block all other clients connected to the storage from doing their regular backups. My backup/prune commands look like this (encryption/password related stuff omitted for brevity): duplicacy-wrapper backup duplicacy-wrapper prune -a -keep 0:365 -keep 30:30 -keep 7:7 -keep 1:1 duplicacy-wrapper prune -a -exclusive duplicacy-wrapper copy -from default -to b2 It is not recommended to run duplicacy prune -exhaustive while there are backups in progress, because new chunks uploaded by these backups may be marked as unreferenced and thus turned into fossils. I'm curious to know how others use prune. I came to the conclusion, under no circumstances a third party (family) would be able to restore my Duplicacy backup in case of need. This may happen with other -keep options. The latest version 2. duplicacy in the repository and put a file named preferences that stores the snapshot id and encryption and storage options. I basically want anything in my cloud storage that has been deleted from my local machine 60 days ago should be removed the cloud storage. I successfully backed up (to B2) several folders. I’m attempting a Prune job on one storage, which contains four snapshot IDs, and I want to avoid pruning two of the IDs. Adjust The issue is, Duplicacy can’t do this because pruning is done on a snapshot level. When in doubt, you can always run the command with -dry-run flag, it will do everything except actually deleting anything. I have been very impressed with Duplicacy. You may be able to find out when this chunk was deleted from these logs files. SNAPSHOT_ID SYNOPSIS: duplicacy copy - Copy snapshots between compatible storages USAGE: duplicacy copy [command options] OPTIONS: -id <snapshot id> copy snapshots with the specified id instead of all snapshot ids -r <revision> [+] copy snapshots with the specified revisions -from <storage name> copy snapshots from the specified storage -to <storage name> copy Yes you need to add check and prune for the B2 storage also (if you want to check and prune the storage), but I’d recommend separating the jobs from the backup schedule - and do the check and prune for both storages in a separate schedule that’s run less frequently than your backups. duplicacy list -files duplicacy prune -r 344-350 (or other revision that I became from the list command) Yes you need to add check and prune for the B2 storage also (if you want to check and prune the storage), but I’d recommend separating the jobs from the backup schedule - and do the check and prune for both storages in a separate schedule that’s run less frequently than your backups. This is from the doc for the check command: This is somehow related with (part of) this other issue, but I’ll open a new thread to keep things focused. The prune command has the task of deleting old/unwanted revisions and unused chunks from a storage. When I finally started running it, it would run for a while, eventually giving an error: “Failed to locate the path for the chunk [] net/http: TLS handshake timeout” I downloaded the latest version, ran a backup, then a prune. I used to run duplicati and this had an export configuration function which mean in case of disaster (OS drive failure) I could get duplicati running on a second machine, import my backup configuration and run a restore (well, the restore part was the part which took for ever but still gchen Sep 8 6:59PM 2017 . This will find all unreferenced chunks and mark them as fossils which will be removed permanently next time the prune command is run with Hi there, I use cli version and my setup is quite straightforward: backup on local storage, copy from local backup to cloud storage and prune identical to the both storages. A check after that will report missing chunks because it doesn’t looking into fossils by default. Duplicacy Web Edition. -all actually means the prune command applies to all ids on that storage. While it does take a little while to get used too since the documentation is a little hit and miss (because as versions and features change, the older documentation doesn’t get updated) I have managed to recover what I needed from backups without any errors. Please describe what you expect to happen (but doesn’t): Would expect Running prune command from /cache/localhost/all Options: [-log prune -storage hidden -a -threads 10 -keep 0:7] 2021-11-17 01:01:33. My questions are: why is there a need for 2 different fossil types, how are they different, and how were they created? How exactly does the duplicacy prune -all command work? Does it completely ignore the snapshot id or does it apply the given retention policy independently to all the snapshots ids?. g. I think @gchen can help better understand what’s going on. prune. Of course first double check the chunk doesn’t exist in the storage, otherwise it is a different issue. You can also recover fossilized chunks with check -fossils -resurrect but since those snapshots were going to be pruned anyway it would be extra work for nothing. I'm currently still doing initial ingestion of my data to B2 and am not running prune commands as of yet. /duplicacy_linux_x64_2. NAME: duplicacy prune - Prune revisions by number, tag, or retention policy USAGE: duplicacy prune [command options] OPTIONS: -id <snapshot id> delete revisions with the specified snapshot ID instead of the default one -all, -a match against all snapshot IDs -r <revision> [+] delete the specified revisions -t <tag> [+] delete revisions with the Prune log files are kept under . With just a few clicks, you can effortlessly set up backup, copy, check, and prune jobs that will reliably protect your data while making the most efficient use of your storage space. It uses Duplicacy under the hood, and therefore supports: Multiple storage backends: S3, Backblaze B2, Hubic, Dropbox, SFTP and more I installed the Duplicacy docker to backup my unraid system to an old synology nas at my mom´s house. I am using Duplicacy web saspus docker image on a Synology nas. I've heard from some that prune is not really useful or necessary and may be potentially problematic. So far I’ve been pruning by date as per the prune documentation (-keep 0:200 -keep 30:60 -keep 7:10 -keep 1:3). Keep the -a argument intact. duplicacy instead (by using -repository when initializing the repository). duplicacy prune [command options] -id <snapshot id> delete revisions with the specified snapshot ID I am using Duplicacy web saspus docker image on a Synology nas. Please describe what you are doing to trigger the bug: duplicacy prune -exhaustive -d with S3 storage driver targeting an idrive e2 bucket. Support. I do not know why my simple mind cannot grasp the prune options but it cannot. duplicacy/logs folder. Then run prune -exhaustive to delete chunks that are not used by any of the remaining snapshots. Hi, I just started trialing Duplicacy and must say that the prune manual read a big ambiguous for me too until I found -keep n:m -> “Keep 1 snapshot every n days for snapshots older than m days. The jobs don't start by default. ”. In your case the storage named b2-trainingandchecking is selected because it is the first one. duplicacy prune -r 1 Deleting snapshot hps_pool2g at revision 1 Fossil collection 1 saved The snapshot hps_pool2g at revision 1 has been removed run ‘duplicacy_linux_x64_3. To clean up storage, run duplicacy prune -exhaustive from another client/repository. Click ok them to edit. The Okay, so this might be an odd case because I didn’t run prune for about a year. 1 prune -keep 0:1 Repository set to / Storage set to b2://mail/ Keep no snapshots older than 1 days Fossil collection 1 found Fossils from collection 1 can't be deleted because deletion criteria aren't met Fossil collection 2 found Fossils from collection 2 can't be deleted because deletion criteria aren Please describe what you are doing to trigger the bug: Running prune -keep 1:1 on snapshot with multiple revisions during the day keeps the oldest revision. Hi guys, I'm new on the scene and I'm just trying to understand how Pruning and Snapshots work. This missing chunk is very likely to be the result of running the prune command with the -exclusive while another backup was running. 128 INFO BACKBLAZE_URL download URL is: https://f003. I have 2x 1 TB Cloud Storage and I had to delete both storages completely and begin from scratch since prune didn’t work and the storage ran out of space after a while. This is the user guide for Duplicacy Web Edition. Yes; you also need either -a flag to prune all snapshot IDs, or specify a specific snapshot ID. If I never prune, all is fine, backups I continue to test Duplicacy via its GUI on a second, Win 7, PC. Please describe what you expect to happen (but doesn’t): The command returns doing nothing. The prune logs have been proven to be very useful when there are missing chunks. You can read more here duplicacy/duplicacy_paper. Here is a sample Hello, I have a couple of questions about how duplicacy prune works cause it is not clear from the --help. With my B2 costs going up I just wanted to limit the amount of backups I keep. So it is clear that I can set all kinds of sophisticated configurations such as duplicacy prune -keep 0:360 -keep 30:180 -keep 7:30 -keep 1:7 which will 1 snapshot per day for Hello, I am trying to grasp the way how duplicacy works. Greyed out. Unfortunately, I can’t check directly on the storage whether the chunks ever existed there, because if they were there, they would have been deleted by that very same prune action (actually, not exactly the one in the OP but I ran the same in exclusive mode shortly after). I wan’t to delete every revision that is older than 30 days. . Because of this, the strategy shown in the documentation for the prune The prune command has the task of deleting old/unwanted revisions and unused chunks from a storage. About e-mail notifications: Is this thread still up to date: Email notification ? Or are there any better options to get a notification only if Remember that prune is a two-step operation: Ref: Lock Free Deduplication · gilbertchen/duplicacy Wiki · GitHub Seems to be something related to the local cache versus the revision in storage. Installation. If you want to move the logs (or the cache) to a different location, you should move . Running out of space on a volume is an event that shall be avoided. You can, for example, run hourly backups and then daily maintenance. There isn't a way to limit the number of backups to keep, but ff you run backup daily, then duplicacy prune -k 0:2 will remove backups older than 2 days thus will likely keep 2 PRUNE_OPTIONS: Set options for each duplicacy prune command, see duplicacy prune command description for details. Topic Replies Views Duplicacy prune -keep-max PR submitted today. duplicacy prune -all (is PC2_user_docs still auto-ignored after 7 days?) It also provide an -ignore option that can be used to skip certain repositories when deciding the deletion criteria. For a test I have installed Duplicacy Web GUI on one PC where I would like to back up 2 internal and 1 external drive which is about 4 TB totally. I run the I'm new on the scene and I'm just trying to understand how Pruning and Snapshots work. Revan335 10 June 2024 05:18 #5. Once you have the prune job added it’s arguments will be displayed in the column in the list. I quit the GUI from the tray icon. The Schedule and Prune settings were not saved. If a chunk reported as missing in fact does not exist in the storage, then you may need to find out why it is missing. The storage was added using add -copy. I have an hourly backup with a prune policy as follows: -keep 0:365 -keep 24:30 -keep 30:7 -a -threads 30 I believe this should result in no revisions older than a year, then two per month (ish), one per day (ish) then one per hour. EDIT: Based on my test, it seems to work per snapshot basis (the other option would not make much sense). 1 Like. The prune command is the only command that can delete chunks, and by default Duplicacy always produces a prune log and saved it under the . 2. This means that -keep 7:90 might be too aggressive for your need. But what about snapshots newer than 30 days? Will they be removed or kept as is? What happens when I run duplicacy prune -keep 0:720 -keep 30:360 -keep 7:30 -keep 1:7 -a -exhaustive Schedule. Reading through the prune help pages and various other posts, I think my settings are correct: -keep 0:14 -keep 7:7 -keep 1:1 Which should translate to: -keep 0:14 # Is snapshot older that 14 days? Yes keep 0 revision Here: prune · gilbertchen/duplicacy Wiki · GitHub. I continually run duplicacy every 6 hours and want to keep everything for the last week (days 0 hi there, I have tried to play around with pruning but I’m afraid I can’t figure out, how it works. By running the check command daily and the prune command weekly or bi-weekly, you ensure backup integrity and efficient storage management. This needs to be said strongly so users doesn’t get surprised: When you start using prune, you will lose some deleted files and original versions of changed files. 37G, but on the First thought was to prune the whole storage and start fresh. When I run duplicacy prune -keep 30:30 then it will remove all snapshots that are older than 30 days except one for each 30days period, sure. I run the backup tasks every day and my goal is to keep the backups of the last 5 or 7 days with no monthly or yearly saved backup (it´s just some private stuff). “Missing chunk”), duplicacy-web still runs the prune task the combination of the above 2 bugs have caused me great pain. 498 INFO RETENTION_POLICY Keep no I’ve read the command docs here: As well as various threads tagged with “prune” which ask similar questions: I believe I understand the algorithm used by the command when executed, however I’m struggling to create a command that reflects what I actually want. Hoping someone can tell me how to set the prune options to achieve what I would like which is very simple. 066 INFO STORAGE_SET Storage set to hidden 2021-11-17 01:01:34. Is it even necessary? It would be great if you can explain it in easy language what it does :). duplicacy is discardable. Backing up locally and to GDrive. Can someone please tell me if I’m right with the following command and expected results? 😅 duplicacy prune -keep 0:364 -keep 24:182 -keep 7:91 -keep 1:1 Keep 1 Getting started Duplicacy licenses (free and paid) Download Duplicacy or Build Duplicacy from source Quick Start tutorial (Web-UI version) Quick Start tutorial (CLI version) Quick Start tutorial (GUI version) Supported storage backends I would appreciate if someone could help me understand prune and what revisions are kept. So I set up prune to run once a week with the following parameters -keep 0:365 -keep 7:30 -keep 1:7 -a. The snapshot id is azinchen/duplicacy is a Docker image created to easily perform automated backups. dsqbfnb zcva hmwr tkausp xnreww jkrky thxq wqkbrl omkv mwuyfbe