AJAX Error Sorry, failed to load required information. Please contact your system administrator. |
||
Close |
Duplicacy prune You can read more here duplicacy/duplicacy_paper. Hi, I just started trialing Duplicacy and must say that the prune manual read a big ambiguous for me too until I found -keep n:m -> “Keep 1 snapshot every n days for snapshots older than m days. About e-mail notifications: Is this thread still up to date: Email notification ? Or are there any better options to get a notification only if Remember that prune is a two-step operation: Ref: Lock Free Deduplication · gilbertchen/duplicacy Wiki · GitHub Seems to be something related to the local cache versus the revision in storage. The prune command has the task of deleting old/unwanted revisions and unused chunks from a storage. For a test I have installed Duplicacy Web GUI on one PC where I would like to back up 2 internal and 1 external drive which is about 4 TB totally. Without prune the datastore is add-only. I run the I'm new on the scene and I'm just trying to understand how Pruning and Snapshots work. prune. I have been very impressed with Duplicacy. Why I can’t Hi, i just performed a backup as a test and later i did duplicacy prune -exclusive -r 1 The revision is removed but in the chunks dir on the storage i still see some empty chunk dirs. Let’s go through this piece by piece. Today I got a warning that I am exceeding my limits. Duplicacy comes with a newly designed web-based GUI that is not only artistically appealing but also functionally powerful. What does Pruning do exactly? The Keep 1 snapshot feature makes sense however duplicacy prune - Prune revisions by number, tag, or retention policy. 498 INFO RETENTION_POLICY Keep no I’ve read the command docs here: As well as various threads tagged with “prune” which ask similar questions: I believe I understand the algorithm used by the command when executed, however I’m struggling to create a command that reflects what I actually want. You can add -exclusive flag to do it quicker but ensure there is no other duplicacy process interacting with the datastore at the same time. Because of this, the strategy shown in the documentation for the prune The prune command has the task of deleting old/unwanted revisions and unused chunks from a storage. Please describe what you expect to happen (but doesn’t): Would expect Running prune command from /cache/localhost/all Options: [-log prune -storage hidden -a -threads 10 -keep 0:7] 2021-11-17 01:01:33. I think @gchen can help better understand what’s going on. But what about snapshots newer than 30 days? Will they be removed or kept as is? What happens when I run duplicacy prune -keep 0:720 -keep 30:360 -keep 7:30 -keep 1:7 -a -exhaustive Schedule. Loses all changes between oldest and youngest revisions. I've heard from some that prune is not really useful or necessary and may be potentially problematic. I find it unacceptable, and until it’s fixed I’m not using prune. SNAPSHOT_ID SYNOPSIS: duplicacy copy - Copy snapshots between compatible storages USAGE: duplicacy copy [command options] OPTIONS: -id <snapshot id> copy snapshots with the specified id instead of all snapshot ids -r <revision> [+] copy snapshots with the specified revisions -from <storage name> copy snapshots from the specified storage -to <storage name> copy Yes you need to add check and prune for the B2 storage also (if you want to check and prune the storage), but I’d recommend separating the jobs from the backup schedule - and do the check and prune for both storages in a separate schedule that’s run less frequently than your backups. The latest version 2. So far I’ve been pruning by date as per the prune documentation (-keep 0:200 -keep 30:60 -keep 7:10 -keep 1:3). The Okay, so this might be an odd case because I didn’t run prune for about a year. Yes. The Schedule and Prune settings were not saved. Here is a sample Hello, I have a couple of questions about how duplicacy prune works cause it is not clear from the --help. Please describe what you expect to happen (but doesn’t): The command returns doing nothing. Which has worked great. The main beef is Pruning. /duplicacy_linux_x64_2. You can use this list · gilbertchen/duplicacy Wiki · GitHub to find out which revisions contain specific file. -all actually means the prune command applies to all ids on that storage. duplicacy list -files duplicacy prune -r 344-350 (or other revision that I became from the list command) Yes you need to add check and prune for the B2 storage also (if you want to check and prune the storage), but I’d recommend separating the jobs from the backup schedule - and do the check and prune for both storages in a separate schedule that’s run less frequently than your backups. Revan335 10 June 2024 05:18 #5. I have only seen it with -keep 1:1. This is the full-featured web-based GUI frontend supporting all major operations (backup, restore, check, copy, and prune) as well as other features such as scheduling and email notifications. I tried again, this time Starting each duplicacy prune -keep 0:360 -keep 30:180 -keep 7:30 -keep 1:7 1 Like. I’m using the following options: -exclusive -keep 0:90 -keep 7:30 -keep 1:3 -ignore clonezilla -ignore scyllams_part The results are that nothing gets pruned in the storage, and I suspect I’m using the “-ignore” option incorrectly. duplicacy prune -r 1 Deleting snapshot hps_pool2g at revision 1 Fossil collection 1 saved The snapshot hps_pool2g at revision 1 has been removed run ‘duplicacy_linux_x64_3. You may be able to find out when this chunk was deleted from these logs files. You can, for example, run hourly backups and then daily maintenance. I have 2x 1 TB Cloud Storage and I had to delete both storages completely and begin from scratch since prune didn’t work and the storage ran out of space after a while. backblazeb2. I have this command running every To summarize: if you did not violate duplicacy assumptions: did not touch the same snapshot ID from two different places, did not use -exclusive flag, did not interrupt prune, there shall be no way to silently lose data. Of course first double check the chunk doesn’t exist in the storage, otherwise it is a different issue. With my B2 costs going up I just wanted to limit the amount of backups I keep. When in doubt, you can always run the command with -dry-run flag, it will do everything except actually deleting anything. I have an hourly backup with a prune policy as follows: -keep 0:365 -keep 24:30 -keep 30:7 -a -threads 30 I believe this should result in no revisions older than a year, then two per month (ish), one per day (ish) then one per hour. Moreover, since most cloud storage services do not provide a locking service, the best effort is to use some basic file Will this work: duplicacy prune -keep 0:20. Please note that if there are multiple Duplicacy instances running on different [root@mail ~]# . duplicacy in the repository and put a file named preferences that stores the snapshot id and encryption and storage options. Installation. In your case the storage named b2-trainingandchecking is selected because it is the first one. Snapshot Pruning. NAME: duplicacy prune - Prune revisions by number, tag, or retention policy USAGE: duplicacy prune [command options] OPTIONS: -id <snapshot id> delete revisions with the specified snapshot ID instead of the default one -all, -a match against all snapshot IDs -r <revision> [+] delete the specified revisions -t <tag> [+] delete revisions with the Prune log files are kept under . Please describe what actually happens (the wrong behaviour): Output of Found redundant chunk floor(# of chunks Delete old snapshots from snapshots sub folder. The add command is similar to the init command, except that the first argument is a storage name used to distinguish different storages: Running the web-based GUI (Duplicacy Web Edition) to restore, check, copy, or prune existing backups You can use the personal GUI license to back up any files if you are a student, or a staff/faculty member at a school or university. Unfortunately, I can’t check directly on the storage whether the chunks ever existed there, because if they were there, they would have been deleted by that very same prune action (actually, not exactly the one in the OP but I ran the same in exclusive mode shortly after). I successfully backed up (to B2) several folders. Reference: Running prune command from /cache/localhost/all Options: [-log prune -storage hidden -a -threads 10 -keep 0:7] 2021-11-17 01:01:33. Most If you’ve already deleted the unused snapshots under the snapshots directory in your storage, then you can use a prune to identify unreferenced chunks. I am totally new to Duplicacy so I am sorry if I am asking an idiotic question that is obvious to all. 1 Like. The storage was added using add -copy. I currently schedule a backup, check and prune daily, but having issues where it seems old revisions aren’t deleting. Prune: Weekly or bi-weekly, based on your data change rate. The jobs don't start by default. On Windows and macOS, Duplicacy Web Edition is provided as installers. Then run prune -exhaustive to delete chunks that are not used by any of the remaining snapshots. Depends on your target storage performance (one drive — slow, Amazon s3 — fast), amount of data turnover Under the hood, it will create a directory named . exe prune -id laptop -keep 365:365 -keep 30:90 -keep 7:10 -keep 1:2 -threads 64 Please describe what you expect to happen (but doesn’t): Delete old snapshot according to the criteria. Duplicacy’s -keep command controls what to keep – e. robert. Your usecase as described is pretty straightforward: init repo, do backup, run check and/or restore. Please describe what actually happens (the wrong behaviour): It writes a long list of Deleting snapshot laptop I’m backing up several machines to my duplicacy repo but some of them somewhat sporadically. Bluebeep Jan 7 1:04PM 2018 GUI. I came to the conclusion, under no circumstances a third party (family) would be able to restore my Duplicacy backup in case of need. Of course it fails. I quit the GUI from the tray icon. You can also recover fossilized chunks with check -fossils -resurrect but since those snapshots were going to be pruned anyway it would be extra work for nothing. This means that -keep 7:90 might be too aggressive for your need. muench 19 February 2019 08:50 #3. The user guide for the old GUI version can be found here. All duplicacy command (except the copy command) only work with one storage at a time. Is it even necessary? It would be great if you can explain it in easy language what it does :). 0 prune -exhaustive’ to remove chunks associated with the now-deleted snapshot, presuming I can also find the preferences that the duplicacy command wants. Hi, looking for a bit of help please. I run the backup tasks every day and my goal is to keep the backups of the last 5 or 7 days with no monthly or yearly saved backup (it´s just some private stuff). My backup/prune commands look like this (encryption/password related stuff omitted for brevity): duplicacy-wrapper backup duplicacy-wrapper prune -a -keep 0:365 -keep 30:30 -keep 7:7 -keep 1:1 duplicacy-wrapper prune -a -exclusive duplicacy-wrapper copy -from default -to b2 It is not recommended to run duplicacy prune -exhaustive while there are backups in progress, because new chunks uploaded by these backups may be marked as unreferenced and thus turned into fossils. Running out of space on a volume is an event that shall be avoided. Il est conçu pour simplifier le processus de sauvegarde tout en optimisant l'espace de stockage en éliminant les données redondantes. This is because for chunks to be deleted, the first prune collects prospective chunks to be deleted, then waits for at least one backup to be completed for all repositories to be sure they’re no longer needed (any that are, Now I want to completely delete all backup revisions of a specific backup id. Keep the -a argument intact. Support. The snapshot id is azinchen/duplicacy is a Docker image created to easily perform automated backups. If you wanted to actually remove the unreferenced chunks, then remove the -dry-run option. I think the appropriate command would be duplicacy prune -all -exhaustive -dry-run. You can select the second one by New Duplicacy user here, running the Web-UI version in a Docker environment from my Synology NAS. However, there is one thing I think should be handled differently: the behaviour of prune -keep . I am using the Web GUI in a Docker environment. The prune command is the only command that can delete chunks, and by default Duplicacy always produces a prune log and saved it under the . Wasabi’s 90-day minimum for stored data means there is no financial incentive to reduce utilization through early pruning of snapshots. Reading through the prune help pages and various other posts, I think my settings are correct: -keep 0:14 -keep 7:7 -keep 1:1 Which should translate to: -keep 0:14 # Is snapshot older that 14 days? Yes keep 0 revision Here: prune · gilbertchen/duplicacy Wiki · GitHub. Click here for a list of related forum topics. I wan’t to delete every revision that is older than 30 days. This appears to have fixed the issue and the entire backup job is now running in 2-4 hours (that includes creating/copying images of two server volumes, backing up the server target volume to the local Duplicacy storage pool, pruning both the local and cloud storage pools, copying the local storage pool to the cloud with snapshots from four targets). Feature. Any and all help is appreciated as this is the first time I’m trying to interact with the CLI to do anything. I'm currently still doing initial ingestion of my data to B2 and am not running prune commands as of yet. link or the plus sign at the top right corner. Ok. I continually run duplicacy every 6 hours and want to keep everything for the last week (days 0 hi there, I have tried to play around with pruning but I’m afraid I can’t figure out, how it works. When I run duplicacy prune -keep 30:30 then it will remove all snapshots that are older than 30 days except one for each 30days period, sure. That’s it, simple. There is safety I have a local and remote storage. RUN_JOB_IMMEDIATELY: Set to yes to run duplicacy backup and/or duplicacy prune command at container startup. It seems it works as designed. All seems to work fine except that the cloud storage grows in size despite pruning. My questions are: why is there a need for 2 different fossil types, how are they different, and how were they created? How exactly does the duplicacy prune -all command work? Does it completely ignore the snapshot id or does it apply the given retention policy independently to all the snapshots ids?. To clean up storage, run duplicacy prune -exhaustive from another client/repository. duplicacy/logs folder. duplicacy-util duplicacy prune -keep 0:360 -keep 30:180 -keep 7:30 -keep 1:7 Second question. Prune options are not set by default. I'm curious to know For those of us with a finite amount of space and a closed wallet (almost everyone), pruning makes complete sense - not least because having too many snapshots, NAME: duplicacy prune - Prune revisions by number, tag, or retention policy USAGE: duplicacy prune [command options] OPTIONS: -id <snapshot id> delete revisions with the specified snapshot ID instead of the default one -all, -a match against all snapshot IDs -r <revision> [+] delete the specified revisions -t <tag> [+] delete revisions with the specified tags Check if the chunk was not deleted by prune. By running the check command daily and the prune command weekly or bi-weekly, you ensure backup integrity and efficient storage management. 1: 270: 10 December 2024 Not able to renew license. Topic Replies Views Duplicacy prune -keep-max PR submitted today. So: duplicacy list Snapshot hps_pool2g revision 1 created at 2017-10-03 16:56 -hash Snapshot hps_pool2g revision 2 created at 2017-10-06 14:08. “I’d definitely not tinker with the /chunks directory if I were you - if you still have any backup data in A prune operation will completely block all other clients connected to the storage from doing their regular backups. The storage list If running prune a second time didn’t work, you may have to run it with the -exclusive flag, making sure no backups are running while it does. duplicacy is discardable. 066 INFO STORAGE_SET Storage set to hidden 2021-11-17 01:01:34. I have read the page on wiki and pdf with detailed explanation of how pruning works. pdf at master · gilbertchen/duplicacy · GitHub but the gist of it is that duplicacy is fully concurrent, it has to support weird scenarios where five machines run backup to a specific storage and seven other machines try to prune it at the same time. I’m trying to figure out Duplicacy’s versioning/retention settings (and their limitations). This is from the doc for the check command: This is somehow related with (part of) this other issue, but I’ll open a new thread to keep things focused. Is this intended? If yes, why? This Empty folders after prune partially answer my question. -keep 30:14 applies to backup snapshots older than 14 days and keeps one version every 30 days. Remember that no other backup can be running when you run this command (due to the use of -exclusive). 37G, but on the First thought was to prune the whole storage and start fresh. After a job has been added, you can set the There are 3 types of Duplicacy licenses: Running the CLI version to restore, check, copy, or prune existing backups on any computer; Running the CLI version to back up any files on a computer where a GUI license is already installed; Running the web-based GUI (Duplicacy Web Edition) to restore, check, copy, or prune existing backups After that, run duplicacy prune -exclusive -exhaustive to clean up the storage. (Duplicacy Web Edition) to restore, check, copy, or prune existing backups So you can set up the backups using the CLI, and still have the pretty restore screen for when you So I’ve created a prune schedule on it with the flag -keep 0:30 The 30 days have passed however and the email alerts I’m getting from Duplicacy are saying: INFO RETENTION_POLICY Keep no snapshots older than 30 days INFO SNAPSHOT_NONE No snapshot to delete So the policy is being applied but nothing is being done? Running prune Add prune job with the desired schedule and ignore retention settings configurable in the dialog. As a test I configured the jobs to backup a user’s folder tree and one critical software specific folder tree. but over the last few days i've had to split out one of the sources to 3 seperate ones. There are 4 types of jobs: backup, check, copy and prune. It has now run for a day and, as expected, Hi, I’m just starting with Duplicacy and would like to hear the best practice for when I need to have a Check & Prune job in my time schedule. Say I run three backups a day and once every week I prune old backups with the following: duplicacy prune -keep 1:2 -all Duplicacy will keep one revision per day for revisions older than 2 days, fine. If the Prune snapshot after first backup checkbox is checked, Duplicacy will run the pruning operation after the first backup of the day or the manual backup. 498 INFO RETENTION_POLICY Keep no I installed the Duplicacy docker to backup my unraid system to an old synology nas at my mom´s house. Although there is supposedly unlimited space available on Google Drive, I would I've been running duplicacy for a while now. com 2021-11-17 01:01:35. Check: Daily, early morning before the first backup. This missing chunk is very likely to be the result of running the prune command with the -exclusive while another backup was running. The important thing is that there are a number of backup ids which i stopped using 2-3 years ago and waited for them to gradually get pruned, You can delete orphaned snapshots manually (from subfolder “snapshots”) on OneDrive, then clear local duplicacy cache, and run prune -exhaustive to cleanup orphaned chunks. Any guidance on how to I continue to test Duplicacy via its GUI on a second, Win 7, PC. Backing up locally and to GDrive. “Missing chunk”), duplicacy-web still runs the prune task the combination of the above 2 bugs have caused me great pain. I am using Duplicacy web saspus docker image on a Synology nas. Quick overview NAME: Then click on the prune options there and replace the -keep arguments with what you actually wanted. This is the user guide for Duplicacy Web Edition. I run a backup and check every night. If you want to move the logs (or the cache) to a different location, you should move . It uses Duplicacy under the hood, and therefore supports: Multiple storage backends: S3, Backblaze B2, Hubic, Dropbox, SFTP and more I installed the Duplicacy docker to backup my unraid system to an old synology nas at my mom´s house. If I never prune, all is fine, backups A place for Duplicacy users and developers. Click ok them to edit. In the GUI I set non-default settings for Rate Limit, Threads, Volume Shadow Copy Schedule, and I turned on Prune Snapshots. This is because for chunks to be deleted, the first prune collects prospective chunks to be deleted, then waits for at least one backup to be completed for all repositories to be sure they’re no longer needed (any that are, . Can someone please tell me if I’m right with the following command and expected results? 😅 duplicacy prune -keep 0:364 -keep 24:182 -keep 7:91 -keep 1:1 Keep 1 Getting started Duplicacy licenses (free and paid) Download Duplicacy or Build Duplicacy from source Quick Start tutorial (Web-UI version) Quick Start tutorial (CLI version) Quick Start tutorial (GUI version) Supported storage backends I would appreciate if someone could help me understand prune and what revisions are kept. The reasons — it leaves datastore in bad state if interrupted. Once you have the prune job added it’s arguments will be displayed in the column in the list. Duplicacy Web Edition. Quick overview NAME: duplicacy prune - Prune revisions by number, I would like to set up Duplicacy on 8 Windows computers (small company) to back them up completely to Google G Suite Drive (business plan). I know there have been other threads about understanding the Prune command, but after searching through them they’re still not resolving my confusion or, rather, not in a way I understand so I would hugely appreciate some direct help! This is the setup: I have With duplicacy in current state of affairs I don’t use prune. And for snapshots newer than 7 days it keeps every snapshot? So I have daily snapshots for 1 week, and 4 Duplicacy est un puissant outil de sauvegarde dans le cloud qui offre des capacités de sauvegarde et de déduplication de données efficaces et sécurisées. This may happen with other -keep options. There isn't a way to limit the number of backups to keep, but ff you run backup daily, then duplicacy prune -k 0:2 will remove backups older than 2 days thus will likely keep 2 PRUNE_OPTIONS: Set options for each duplicacy prune command, see duplicacy prune command description for details. So it is clear that I can set all kinds of sophisticated configurations such as duplicacy prune -keep 0:360 -keep 30:180 -keep 7:30 -keep 1:7 which will 1 snapshot per day for Hello, I am trying to grasp the way how duplicacy works. I downloaded and installed the Dublicacy Web Edition on a Windows x64 machine and set up a couple of backup jobs. Thanks! I take that as a “yes” regarding the last question . 2. Greyed out. Quick overview NAME: duplicacy prune - Prune snapshots by revision, tag, or retention policy USAGE: duplicacy prune [command options] OPTIONS: -id <snapshot id> delete snapshots with the specified id instead of the default one -all, -a match against all snapshot IDs -r <revision> [+] delete -all doesn’t mean to prune all the storages in the preferences file. duplicacy instead (by using -repository when initializing the repository). Yes, you can remove the subdirectory under the snapshots directory, but this will leave many unreferenced chunks on the storage. On local storage the backup size is ca. But I am still unable to quite fully understand how it works and how is it able to work lock-free. Yes; you also need either -a flag to prune all snapshot IDs, or specify a specific snapshot ID. So I set up prune to run once a week with the following parameters -keep 0:365 -keep 7:30 -keep 1:7 -a. duplicacy prune [command options] -id <snapshot id> delete revisions with the specified snapshot ID I am using Duplicacy web saspus docker image on a Synology nas. They are combinable – from oldest to youngest: duplicacy prune -keep 0:360 -keep 30:180 -keep 7:30 $ duplicacy prune # Chunks from deleted backups will be removed if deletion criteria are met To back up to multiple storages, use the add command to add a new storage. So far everything works great but I am confused with the prune settings. duplicacy prune -all (is PC2_user_docs still auto-ignored after 7 days?) It also provide an -ignore option that can be used to skip certain repositories when deciding the deletion criteria. Please describe what you are doing to trigger the bug: duplicacy prune -exhaustive -d with S3 storage driver targeting an idrive e2 bucket. Hi guys, I'm new on the scene and I'm just trying to understand how Pruning and Snapshots work. While it does take a little while to get used too since the documentation is a little hit and miss (because as versions and features change, the older documentation doesn’t get updated) I have managed to recover what I needed from backups without any errors. If I never prune, all is fine, backups I continue to test Duplicacy via its GUI on a second, Win 7, PC. I basically want anything in my cloud storage that has been deleted from my local machine 60 days ago should be removed the cloud storage. I’m attempting a Prune job on one storage, which contains four snapshot IDs, and I want to avoid pruning two of the IDs. I have the following options enabled on a prune command set to run after every scheduled backup: -d | -keep 7:7 -keep 0:30 -a -exhaustive -threads 10 I interpret this as follows: Keep one revision every 7 days after seven Please describe what you are doing to trigger the bug: I run duplicacy. But for some machines I feel like it would be more suitable to simply keep, for instance, the last 3 snapshots, rather than consider when those I have been using duplicacy for a week now It is absolutely amazing. 1 prune -keep 0:1 Repository set to / Storage set to b2://mail/ Keep no snapshots older than 1 days Fossil collection 1 found Fossils from collection 1 can't be deleted because deletion criteria aren't met Fossil collection 2 found Fossils from collection 2 can't be deleted because deletion criteria aren Please describe what you are doing to trigger the bug: Running prune -keep 1:1 on snapshot with multiple revisions during the day keeps the oldest revision. When I finally started running it, it would run for a while, eventually giving an error: “Failed to locate the path for the chunk [] net/http: TLS handshake timeout” I downloaded the latest version, ran a backup, then a prune. g. 1 seems to have a less verbose Everything under . my checks kept failing for months and I never got any emails, and the prunes kept I have roughly 350 revisions. There’s always something failing. Adjust The issue is, Duplicacy can’t do this because pruning is done on a snapshot level. 1. If a chunk reported as missing in fact does not exist in the storage, then you may need to find out why it is missing. With just a few clicks, you can effortlessly set up backup, copy, check, and prune jobs that will reliably protect your data while making the most efficient use of your storage space. This will find all unreferenced chunks and mark them as fossils which will be removed permanently next time the prune command is run with Hi there, I use cli version and my setup is quite straightforward: backup on local storage, copy from local backup to cloud storage and prune identical to the both storages. Explanation given is that prune deletes files in two steps: first renaming chunks to fossils, then actually removing only if there For this my goal is to run duplicacy every hour, have it email me any errors, and then prune (thin) the backed up data. duplicacy/logs. Duplicacy offre des fonctionnalités telles que la gestion des versions, le SYNOPSIS: duplicacy check - Check the integrity of snapshots USAGE: duplicacy check [command options] OPTIONS: -all, -a check snapshots with any id -id <snapshot id> check snapshots with the specified id rather than the default one -r <revision> [+] the revision number of the snapshot -t <tag> check snapshots with the specified tag -fossils search fossils if a chunk If running prune a second time didn’t work, you may have to run it with the -exclusive flag, making sure no backups are running while it does. Data is not lost — but checks will start failing on the ghost snapshots that were supposed to be deleted. This needs to be said strongly so users doesn’t get surprised: When you start using prune, you will lose some deleted files and original versions of changed files. EDIT: Based on my test, it seems to work per snapshot basis (the other option would not make much sense). Understanding Snapshots and Pruning. This sounds like a good idea if I wanted to prune stale or deprecated repos manually. Hoping someone can tell me how to set the prune options to achieve what I would like which is very simple. ”. Replace delete all -keep arguments and replace them with your own -keep 0:2. A check after that will report missing chunks because it doesn’t looking into fossils by default. It also integrates a dashboard to provide a graphical view of the status of storages and backups. I do not know why my simple mind cannot grasp the prune options but it cannot. . 128 INFO BACKBLAZE_URL download URL is: https://f003. 3: 63: 10 December 2024 Weekly / monthly scheduled tasks in duplicacy_web. I used to run duplicati and this had an export configuration function which mean in case of disaster (OS drive failure) I could get duplicati running on a second machine, import my backup configuration and run a restore (well, the restore part was the part which took for ever but still gchen Sep 8 6:59PM 2017 . I'm curious to know how others use prune. Reading through the prune help pages and various other posts, I think my settings are correct: -keep 0:14 -keep 7:7 -keep 1:1 Which should translate to: -keep 0:14 # Is snapshot older that 14 days? Yes keep 0 revision List of storage names to prune for duplicacy prune operations* check: List of storage names to check for duplicacy check operations* Note that * denotes that this section is mandatory and MUST be specified in the configuration file. Prune ran and completed jus I have been very impressed with Duplicacy. The prune logs have been proven to be very useful when there are missing chunks. The prune command is sent via WebUI: “-id S_Bilder -r 1-1000 -exclusive” (1) duplicacy-web does not send an email on failure for a check task (2) if a schedule has a check task and a prune task, and if the check task fails (e. govdme luh cwvellg djwnz tyryhn wusdrj mvytsb jdya ccd abwti