Jump to content

MandalorePatriot

Members
  • Posts

    8
  • Joined

  • Last visited

Posts posted by MandalorePatriot

  1. Hello,

    During setup of CloudDrive, I made the mistake of not making duplicates of the data I was uploading. So now some of my data is only stored in the cloud and I would like to correct this.

    What is the best way to download the data but still keep it in the cloud as well? I've tried using File Explorer and copying the data to another drive, but the performance drops to 0 until more data is "downloaded". I am assuming it is because the download speed cannot keep up with the copy speed.

    I currently have gig download speeds, is there fine-tuning of the performance that can be done? Thanks in advance for any help!

    -MandalorePatriot

  2. 18 hours ago, Christopher (Drashna) said:

    There is a "clear cache" option that should try to forcibly prune the cache.  Try that, and see if the issue comes back. 

    Apologies, I tried that but nothing is removed. Rebooting also does not clear any of the space and there are no active transfers (just the upload).

  3. So after doing some digging, I think this may be related to active usage of the CloudDrive. I have CloudDrive as a member of my DrivePool, and I point Plex to the DrivePool. While media is being watched, CloudDrive downloads data to the cache. My assumption is that that "downloaded" data is not clearing automatically, because the CloudPart folder keeps increasing. I've tried to clear the cache, but it keeps growing no matter what. Any ideas?

    ***EDIT***
    Just to ensure it is not part of the issue, I removed the cache drive from the DrivePool. I was originally storing some data on it, but because of this I moved non-cloud data off and removed the drive from the pool.

  4. Hello,

    I've been uploading my data for Google Drive without issue, but recently the drive I've been using for cache is filling up. When I first set this up, I set an Expandable cache of 50GB on an 8TB drive. The cache would fill as I moved data, and the To Upload would show the amount to upload. As the To Upload decreased (from uploading), the cache would gradually go down until resting near 50GB again.

    Since I'm using an 8TB drive, I decided to move a large amount of data and let it upload. Recently, the Cached section keeps increasing, even while the To Upload goes down. I included a screenshot. During this time, the drive usage keeps going higher and I am now at about 93% usage. Even with uploading, the drive usage keeps getting higher without any new files being sent.

    Any idea what is causing this? I do use DrivePool and the CloudDrive is a member, but I restrict what data is stored where. I have not changed any DrivePool or CloudDrive settings before this happened, but I am thinking it might be the 90% store on other drives rule in DrivePool.

    Thanks in advance for the help!

    -MandalorePatriot

    CloudDrive.png

  5. 18 hours ago, srcrist said:

    Out of curiosity, does Google set different limits for the upload and download threads in the API? I've always assumed that since I see throttling around 12-15 threads in one direction, that the total number of threads in both directions needed to be less than that. Are you saying it should be fine with 10 in each direction even though 20 in one direction would get throttled?

    What is even the gain of using so many threads? More connections, sure, but doesn't Google throttle bandwidth after a certain amount? And it also depends on your upload speed, I'm capped by IPS @ 40Mbps roughly, so it seems only 3 upload threads is plenty.

  6. 21 hours ago, Christopher (Drashna) said:

    That depends ENTIRELY on your use case.  It's not a question that others can really answer. 

     

    But if performance is important, then the SSD is going to be the better choice for you. 

    But if you're accessing a lot of data (reading and writing), then a hard drive may be a better option.

    For a homelab use, I can't really see reading and writing affecting the SSDs that much. I have an SSD that is being used for firewall/IPS logging and it's been in use every day for the past few years. No SMART errors and expected life is still at 99%. I can't really see more usage in a homelab than that.

    In an enterprise environment, sure, lots of big databases and constant access/changes/etc.

    I have a spare 500GB SSD I will be using for the CloudDrive and downloader cache. Thanks for the responses again everyone! -MandalorePatriot

  7. 3 hours ago, srcrist said:

    SSD. Disk usage for the cache, particularly with a larger drive, can be heavy. I always suggest an SSD cache drive. You'll definitely notice a significant impact. Aside from upload space, most drives don't need or generally benefit from a cache larger than 50-100GB or so. You'll definitely get diminishing returns with anything larger than that. So speed is far more important than size. 

    Thank you, I really appreciate your quick and informative answer!

  8. Hi everyone,

    I looked around on the forum and searched the forum & Google but was not able to find an answer to this question. Which is better for CloudDrive, an SSD or HDD cache? SSD will usually provide better performance with lower space and an HDD will provide more space with lower performance. So what is more important for the cache, space or performance?

    Thanks for reading! -MandalorePatriot

×
×
  • Create New...