Does Dropbox have a file size upload limit? -Chris-
On Mon, 4 Dec 2023, 22:23 MylesDearBusiness via Bacula-users, < bacula-users@lists.sourceforge.net> wrote: > > > Ok, here goes ... > > > root@c1:~# find / -path /mnt -prune -o -type f -print | grep "Vol-0" > root@c1:~# > > > root@c1:~# df -h > Filesystem Size Used Avail Use% Mounted on > udev 941M 0 941M 0% /dev > tmpfs 198M 1.6M 196M 1% /run > /dev/vda1 49G 19G 30G 39% / > tmpfs 986M 20K 986M 1% /dev/shm > tmpfs 5.0M 0 5.0M 0% /run/lock > tmpfs 986M 0 986M 0% /sys/fs/cgroup > /dev/loop0 9.7M 9.7M 0 100% /snap/canonical-livepatch/246 > /dev/loop1 9.9M 9.9M 0 100% /snap/canonical-livepatch/248 > /dev/loop2 74M 74M 0 100% /snap/core22/864 > /dev/loop3 43M 43M 0 100% /snap/doctl/1402 > /dev/loop4 106M 106M 0 100% /snap/core/16091 > /dev/loop5 92M 92M 0 100% /snap/lxd/24061 > /dev/loop6 64M 64M 0 100% /snap/core20/1974 > /dev/loop7 43M 43M 0 100% /snap/doctl/1445 > /dev/vda15 105M 6.1M 99M 6% /boot/efi > /dev/loop8 41M 41M 0 100% /snap/snapd/20092 > /dev/loop9 68M 68M 0 100% /snap/lxd/22753 > /dev/loop10 106M 106M 0 100% /snap/core/16202 > /dev/loop12 41M 41M 0 100% /snap/snapd/20290 > /dev/loop11 2.1G 188K 2.0G 1% /tmp > /dev/loop13 64M 64M 0 100% /snap/core20/2015 > tmpfs 198M 0 198M 0% /run/user/1000 > MylesDearDropBox: 2.1T 651G 1.4T 32% /mnt/MylesDearDropBox > root@c1:~# > > root@c1:~# find /mnt/MylesDearDropBox/Backup/bacula/archive/ > /mnt/MylesDearDropBox/Backup/bacula/archive/ > /mnt/MylesDearDropBox/Backup/bacula/archive/MylesMpwrware1 > /mnt/MylesDearDropBox/Backup/bacula/archive/MylesMpwrware1.l > /mnt/MylesDearDropBox/Backup/bacula/archive/MylesMpwrware2 > root@c1:~# > > > I searched through the entire Dropbox directory and no files with pattern > "Vol-" were found. > > > Best, > > <Myles> > > On 2023-12-04 4:14 p.m., Rob Gerber wrote: > > Maybe Dropbox or rclone or some combination of the two are limiting > > you to 1GiB file sizes? > > In fact, for your rclone process I see it has a > 1GB cache size limit. > " --vfs-cache-max-size 1G" I bet in the case of the > dd command you > did, we filled the write cache and then dd exited. If the > cache was > larger or the input command was rate limited, we might not have > that > issue. Maybe if bacula backups took longer in some cases you > wouldn't > run into this problem. Bacula does have a bandwidth rate limit > > feature, but I'd work on the cache size or a more graceful failure > mode > first (like "is cache full? Make bacula wait a while" - > admittedly > something I don't know to be possible). > > > By default, Bacula's file > based backup writes to file volumes. Think > of it as being like writing > the backed up field and directories to > tar or zip files - the files and > directories that are backed up by > bacula are stored in single large > archive files, using Bacula's own > file format. Conceptually, Bacula isn't > using the tar format, but the > bacula file writers (by default) are using > something like a tar file. > I think it's done this way because way back > when Kern started > developing bacula, the original destination for the > backups was a > tape drive. Later hard drives became cheaper and Kern > realized that > bacula could also write to "file volumes" that were stored > on a hard > drive. This means some customers who couldn't afford a tape > drive but > could afford a larger hard drive could use bacula. Bill > recently > mentioned that bacula can write backed up files and directories > to > some cloud storage solutions directly, so I think the file volume > > method isn't used in every case by bacula, but without special > > configuration on your part bacula is probably using these file > volumes as > described above. > > Please do the following, probably as root: sudo find / > |grep -i > Vol-0 > > Also please do find > /mnt/MylesDearDropBox/Backup/bacula/archive/ > > and please do df -h > > > Robert Gerber 402-237-8692 r...@craeon.net <mailto:r...@craeon.net> > <r...@craeon.net> > > On Mon, Dec 4, 2023, 1:58 PM MylesDearBusiness > <md...@mpwrware.ca > > <md...@mpwrware.ca> <mailto:md...@mpwrware.ca> <md...@mpwrware.ca>> > wrote: > > Hi, Rob, > > Thanks for the response. > > 1. I'm only using 25% > of my 2TB Dropbox account, so I don't expect > storage to be full. > > This > particular cloud server is tiny, just a single CPU, 50GB > storage, 2GB > RAM. > > The biggest file I managed to write successfully to my > > rclone/Dropbox mount is 1GB: > > When I tried to write a bigger file, I got > an "out of memory" error, > in hindsight I suppose this was to be expected. > I'm trying to keep > costs down by renting only a very small cloud machine > until such time > I need the capacity increase. > > root@c1:~# dd > if=/dev/urandom > of=/mnt/MylesDearDropBox/Backup/someuniquefilename.img > bs=1G count=1 > 1+0 records in 1+0 records out 1073741824 bytes (1.1 GB, > 1.0 GiB) > copied, 9.12953 s, 118 MB/s root@c1:~# ls -l > > /mnt/MylesDearDropBox/Backup/someuniquefilename.img -rw-r--r-- 1 root > > root 1073741824 Dec 4 19:31 > > /mnt/MylesDearDropBox/Backup/someuniquefilename.img root@c1:~# > > > so > I'll tune down my bacula director config for max file size of 1G. > > 2. > I'm still confused by what exactly "Vol-xxx" is supposed to be, I > see > there are config settings for setting this name, but I only > create the > device files MylesMpwrware<x> and point to them in in > bacula-sd > configuration as "Archive Device". Should I also be > creating the > "Vol-xxx" files as well? I did see the first of my > "Archive Device" files > filling up: > > root@c1:~# ls -l > /mnt/MylesDearDropBox/Backup/bacula/archive/ total > 20971544 -rw-r--r-- 1 > root root 21474860756 Dec 4 03:27 > MylesMpwrware1 -rw-r--r-- 1 root root 0 > Dec 4 03:04 > MylesMpwrware1.l -rw-r--r-- 1 root root 0 Dec 4 01:00 > > MylesMpwrware2 root@c1:~# > > > I'm sure with a little more banging my > head against the wall things > will start to make sense. > > Thanks, > > > <Myles> > > On 2023-12-04 2:26 p.m., Rob Gerber wrote: >> dd > if=/dev/urandom >> of=/mnt/yourdropboxmountpoint/someuniquefilename.img > bs=50G >> count=1 > > _______________________________________________ > Bacula-users mailing list > Bacula-users@lists.sourceforge.net > https://lists.sourceforge.net/lists/listinfo/bacula-users >
_______________________________________________ Bacula-users mailing list Bacula-users@lists.sourceforge.net https://lists.sourceforge.net/lists/listinfo/bacula-users