Comments

  • How exactly does an incremental backup work?
    When files get deleted on the source, they are flagged to be purged after the period of time set in the backup plan for keeping deleted files for x days/ weeks/ months.
    Incrementals backup new and modified files. The weekly full backs up all files that have outstanding block level incremental versions.
    Your full Quickbooks file for example might be 300MB. The daily incremental is probably around 10MB.
    When you run the next full, it backs up the entire 300MB.
  • Backup Agent 7.8 for Windows / Management Console 6.3
    The way I see it, they are mutually exclusive. The only value of FFI is to those of us who do not use GFS or immutability.
    It will hopefully make NBF file backups have similar storage consumption as Legacy format has.
    Am still trying to fully understand the new features, but when I do, I will write-up my assessment in this forum.
  • Bandwidth Settings
    I use the MSP version of the SW but I believe the code is the same.
    I set "When using Cloud/Local storage" to unlimited.
    I then specify the workday hours cloud bandwidth to something less as you have done, but leave local at unlimited since it has no impact on the performance of anything.
    During the workday the screen shows "specified" - and when I am in off-hours it shows Unlimited.
    I have not tested to see if perhaps the software is not using local time to adjust the bandwith.
    The most important thing is whether it is actually lowering your bandwidth during the peak hours - something that you can see when the plan is running.
  • "retrieving data" operation takes a lot of time, is it normal? How can I speed it up?
    The new format does not support incremental forever, so it is not a good fit for Cloud file backups as it required periodic "Fulls" which require re-uploading all 2TB of data.
    For Image and VHD CLoud backups - and for local drives - the new format works great. You would just need a larger NAS or USB drive.
    You could run a FULL once a month and incrementals each night.
    If you want to keep 90 days worth of backups of 2TB of data on a local drive, you will consume 8TB plus the daily incremental space consumption.
    I would recommend at least 10TB of backup drive space, preferrably12TB.
    I also recommend compression and encryption. They do not slow down the backups unless you are on a horrendoesly old slow machine.
    I would also encourage you to utilize Wasabi or Backblaze to keep a copy of the data using legacy mode incremental forever in the cloud as a local NAS device can break, get stolen (and not encrypting the data is asking for touble), or be destroyed in a disaster.
    Alex is correct, the new format is orders of magnitude faster in bringing up the contents of storage to view/restore than legacy mode for large file stores, but takes significantly more space.
    Local space is a one-time and relatively cheap.
    Hope this helps.
    - Steve
  • How to Restore to Azure VM
    HYPERV replication does not rely on backups. It is a windows feature that mirrors the state of the primary server and can be configured to keep multiple checkpoints to restart from on the HYPERV replication server. It is a manual failover - but can be completed in minutes.
    We use ramnode to host several clients' servers, as well as to spin up an instance in a disaster scenario where using our local spare server is not a viable option.
    Ramnode is significantly cheaper than Azure or AWS and has better, more accessible, technical support. We have used all three and prefer Ramnode overall.
    We only use he HYPERV replication feature for our largest, most critical customers, where downtime due to hardware or OS failures cannot exceed an hour.
  • Immutable Backups
    As for your dilemma, what you really want (and unfortunately cannot have) is the legacy format sent to Immutable storage.
    Why do you need immutable storage?
    Is the customer demanding that? Or are you just trying to us Immutability to protect your data?
  • How to Restore to Azure VM
    HyperV replication is what we use for clients that need fast recovery from major, non-Disaster events.
    We keep the old server when upgrading and make it the replica.
    Other MSP’s rely on Datto-type devices to spin up virtual machines on prem.
    Neither of these approaches help in a true Disaster where the client equipment and/or site is unusable for whatever reason.
    For that we rely on a spare server in our office that can quickly download the client’s image/vhdx file(s), or we spin-up a Ramnode hosted instance and do the recovery from there. We tell clients that in a true disaster scenario, we will aim to have core systems up and running in less than 24 hours from the time a disaster is declared, with data from the previous backup.
  • How to Restore to Azure VM
    We spun up a Ramnode instance and did a VHDx restore from Google Nearline. We achieved an average download speed roughly equivalent to a local USB drive - around 20MB/sec, 1.2 GB/minute. At that rate, a 100GB restore would take about 1.5 hours.
    Your mileage may vary.
  • Bad Gateway and 3015 internal storage error
    I am told by support that they have fixed the new Access/Refresh token authorization system. So far today have not received any more of these errors.
  • Issues with Cloud Backups
    Everyone’s requirements and preferences are different, but I can share with you what we do for clients with slow upstream connection speed ( anything under 15mbps).
    We send nightly legacy-format file backups to the local hard drive, and also to Amazon (ZIA).
    We keep the local file backups for a year and the cloud backups for three months.
    We also do weekly new format Image or VHDx backups, using the synthetic full feature which speeds up the weekly backup considerably. Once you complete the first full backup, the subsequent synthetic fulls take 1/6 the time on average.
    Wasabi, BackBlaze and Amazon support synthetic fulls.
    We do legacy format local image/VHDx backups to the USB drive, a full each week or month, and incrementals each night.
    We also do a separate local file backup and call it “recent files”. This runs every couple of hours during the workday to provide intra-day recovery, but we only keep these files for a month or so.
    New backup format is the way to go for all cloud Image or VHDx backups, but for everything else, we use the legacy format as it utilizes far less space in the cloud.
    Hope this helps.
  • Adding a new storage to multiple companies
    When you create a new storage destination, you can choose what companies get access to it, but I do not know of a way to assign an existing destination to multiple companies.
  • image restore verification
    Did you select a new format backup? It does not work with legacy format plans. Once you select the new format, each plan will have a panel where you can select restore verification.
  • Legacy and new format backup
    You are correct about the lifecycle policy behavior, I was asking if MSP360 backups would recognize the contents of the cbl files , old versions and all, such that I could “seed” 90 days worth of backups to the cloud that were originally on the local drive.
  • Legacy and new format backup
    They are anywhere from 3-17GB each.
    So this leads to my next question:
    Lets say i have files with 90 days worth of versions backed up in these CBL files (new format) on my local drive. If i were to copy those files to Amazon using CB Explorer into the appropriate MBS-xxxxx folder, would a subsequent repo sync see them and allow me to "continue on" with a new format Cloud Backup?
    Assuming that it would work, if i were to then use Bucket Lifecycle management to migrate CBL files to a lower tier of cloud storage, would every file contained in the CBL have to match the aging criteria?
    Or is there some intelligence as to what gets put into each CBL file?
  • Legacy and new format backup
    Thanks David. I will do a test tomorrow.
  • Immutable Backups
    The strategy we use to protect our local backups.
    - Put the backup USB HD on the hyperV host and then share it out to the guest VM’s.
    - Use a different admin password on the hyperV than is used on the guest VM’s ( in case someone gets into the file server guest vm where all of the data resides).
    - Do not map drives to the backup drive.
    - Use the agent console password protection feature including protecting the CLI).
    - Turn off the ability to delete backup data and modify plans from the agent console ( company: custom options). You can always turn it on/ back off as necessary to do modifications.)
    - Encrypt the local backups as well somthat if someone walks off with the drive it ismof no use to them.
    We also do image/ VHDx backups to the cloud and file backups to not one, but two different public cloud platforms.
  • 3-2-1 with the new backup format
    In a word, yes. It is a safe and effective way. It allows for flexibility in retention and in file format.
  • File Server best practices
    Couple of questions:
    - Where are you sending the backups to?
    - What are the ISP uplink speeds of your clients?
    - Are you using the HyperV version of the software? Or just using the file backup version?

    We have several clients with large data-vhdx files. We run weekly synthetic fulls to Backblaze B2 and incrementals the other six nights.
    The synthetic fulls take at most 10-12 hrs over the weekend. The incrementals are fast (1-2 hrs.)
    But we have two clients with large data files and painfully slow upload speeds (5 mbps). For those it is simply not feasible to complete even a weekly backup over the weekend. We de-select the data vhdx for those clients.
    Doing synthetic fulls each week will reduce the amount of time it takes as there are fewer changed blocks to upload compared to once per month.
    We have been pushing clients to uplaod to higher speed plans/ fiber as it makes disaster recovery. Lot faster when we can essentially bring back their entire system from the previous day in one set of (large) VHDx restores.
    If slow uplinks are the issue and they cannot be increased to say 25 up, then only upload the non- data VHDx’s to the cloud and rely on the file- level ( old format) for data recovery. I trust that you are doing file-level backups to tne cloud anyway so as to have 30/90 days version/ deleted file retention.
  • Legacy and new format backup
    David- Great write-up explaining the differences. Can you help me get a feel for the reduction in object count using the new format? I use Amazon lifecycle mgmt to move files from S3 to Archive after 120 days ( the files that do not get purged after my 90 day retention expires). However, the cost of the API calls makes that a bad strategy for customers with lots of very small files ( I’m talking a million files that take up 200GB total). If I were to reupload the files in the new format to Amazon and do a weekly synthetic full, ( such that I only have two fulls for a day or so then back to one), would the objects be significantly larger such that the Glacier migration would be cost-effective?
  • Connection error Code: 1056
    Since upgrading to the 7.7.0.107 release I have had seven machines give me license errors or "an unexpected error has occurred". I worked around the license errors by switching, then switching back, the license type. Just installed the .109 release and it fixed the unexpected error issue.