Jump to content
Covecube Inc.
  • Announcements

    • Christopher (Drashna)

      Login issues   11/07/17

      If you have issues with logging in, make sure you use your display name and not the "username" or email.  Or head here for more info.   http://community.covecube.com/index.php?/topic/3252-login-issues/  
    • Christopher (Drashna)

      Getting Help   11/07/17

      If you're experiencing problems with the software, the best way to get ahold of us is to head to https://stablebit.com/Contact, especially if this is a licensing issue.    Issues submitted there are checked first, and handled more aggressively. So, especially if the problem is urgent, please head over there first. 
  • 0
Mirabis

DrivePool + CloudDrive Setup Questions

Question

Hello,

I'm using Windows Server 2016 TP5 (Upgraded from 2012R2 Datacenter..for containers....) and have been trying to convert my Storage Spaces to StableBit Pools. So far so good, but I'm having a bit of an issue with the Cloud Drive.

 

Current:
- Use SSD Optimizer to write to one of the 8 SSDs (2x 240GB / 5x 64GB) and then offload to one of my harddisks ( 6x WD Red 3TB / 4x WD Red 4 TB).
- I've set balancing to percentage (as the disks are different size)

- 1x 64GB SSD dedicated to Local Cache for Google Drive Mount (Unlimited size / specified 20TB)

 

Problem 1:
I've set my Hyper-V folder to duplicate [3x] so I can keep 1 file on SSD, 1 on HDD and 1 on Cloud Drive... but it is loading from CLoud Drive only. This obviously doesn't work as it tries to stream the .vhd from the cloud.

Any way to have it read from the ssd/local disk, and just have the CloudDrive as backup?

Problem 2:
Once the CacheDisk fills up everything slows down to a crowl..... any way I can have it fill up an HDD after the ssd so other transfers can continue? After which it re-balances that data off?

 

Problem 3:

During large file transfers the system becomes unresponsive, and at times even crashes. I've tried using Teracopy (which doesn't seem to fill the 'modified' RAM cache, but is only 20% slower... = less crashes.... but system still unresponsive.

 

Problem 4:

I'm having

  • I/O Error: Trouble downloading data from Google Drive.
  • I/O Error: Thread was being aborted.
  • The requested mime type change is forbidden (this error has occurred 101 times).

Causing the Google Drive uploads to halt from time to time. I found a solution on the forum (manually deleting the chunks that are stuck). But instead of deleting I moved them to the root, so they could be analysed later on (if neccesary).

 

Problem 5 / Question 1:
I have Amazon Unlimited Cloud Drive, but it's still an experimental provider. I've tried it and had a lot of lockups/crashes and an average of 10mbits upload - so I removed it. Can I re-enable it once it exists experimental and allow the data from the Google Drive to be balanced out to Amazon Cloud Drive (essentially migrating/duplicating to the other cloud)?
 

Question 2:
Does Google Drive require Upload Verification? Couldn't find any best practices/guidelines on the settings per provider.
 

 

Settings Screenshots:

 

 

Share this post


Link to post
Share on other sites

8 answers to this question

Recommended Posts

  • 0

What version of StableBit CloudDrive are you using?

 

 

 

Problem 1:
I've set my Hyper-V folder to duplicate [3x] so I can keep 1 file on SSD, 1 on HDD and 1 on Cloud Drive... but it is loading from CLoud Drive only. This obviously doesn't work as it tries to stream the .vhd from the cloud.

Any way to have it read from the ssd/local disk, and just have the CloudDrive as backup?

 
Is read striping enabled? 
 
Also, are you using the UI to confirm access? Or using something like RESMON? 
 
As for forcing the CloudDrive as a backup? Not currently. Having it outside of the pool would be the best/only choice for t his, right now. 
 
 
 

Problem 2:
Once the CacheDisk fills up everything slows down to a crowl..... any way I can have it fill up an HDD after the ssd so other transfers can continue? After which it re-balances that data off?

 

Not currently. And the slowdown is by design, to prevent serious issues from occurring (Windows has some really odd/bad behavior when waiting on IO to complete, slowing it down is a good way to help prevent this from occurring). 

 

I've added a request for this, but it may not happen, and if it does ... it would be a long way off before it got implemented.

https://stablebit.com/Admin/IssueAnalysis/26048

 

 

Problem 3:
During large file transfers the system becomes unresponsive, and at times even crashes. I've tried using Teracopy (which doesn't seem to fill the 'modified' RAM cache, but is only 20% slower... = less crashes.... but system still unresponsive.

 

By crash, you mean BSOD? If so, please grab the memory dumps.

http://wiki.covecube.com/StableBit_CloudDrive_System_Crashes

 

And if you're not already, could you see if the 1.0.0.592 works better? 

http://dl.covecube.com/CloudDriveWindows/beta/download/StableBit.CloudDrive_1.0.0.592_x64_BETA.exe

 

Problem 4:
I'm having

  • I/O Error: Trouble downloading data from Google Drive.
  • I/O Error: Thread was being aborted.
  • The requested mime type change is forbidden (this error has occurred 101 times).
Causing the Google Drive uploads to halt from time to time. I found a solution on the forum (manually deleting the chunks that are stuck). But instead of deleting I moved them to the root, so they could be analysed later on (if necessary).

 

 

Ah, you're using an older version, are you not? 

1.0.0.463, correct? 

 

If so, this is due to a weird API issue (meaning, it's one that Google didn't actually include in their well written documentation, and no, there is no sarcasm here, actually).  Specifically, we cannot query the same drive more than 20 times in a short period of time (seconds, IIRC).  It's a known issue and it's been addressed in the newer beta builds. 

 

However, it may require you to recreate the drive, as it's partially an issue with the "format" that the raw data is using. 

 

Problem 5 / Question 1:
I have Amazon Unlimited Cloud Drive, but it's still an experimental provider. I've tried it and had a lot of lockups/crashes and an average of 10mbits upload - so I removed it. Can I re-enable it once it exists experimental and allow the data from the Google Drive to be balanced out to Amazon Cloud Drive (essentially migrating/duplicating to the other cloud)?

 

As for Amazon Cloud Drive, again, try the newer build. There are a lot of stability, reliability and performance fixes.  It may cause the Amazon Cloud Drive provider to run better. 

 

And as for re-enabling it, with the newer version, you should be safer to do so. Though, I'd recommend making sure that you use different drives for the cache, to optimize performance. 

 

Question 2:
Does Google Drive require Upload Verification? Couldn't find any best practices/guidelines on the settings per provider.

 

For Google Drive, it may not be necessary.  Some providers, like Amazon Cloud Drive, it's essentially required (it loses data if you don't verify, but if you always verify, it seems to be fine ....)  

 

Basically, it changes the upload process. Instead of clearing out the update data, it downloads it, and verifies that it uploaded properly.  If it did, it clears the data. But if it doesn't, it will reupload it. 

 

If you have the bandwidth (500/500, mbps, I'm assuming), then it's worth having on, as it's another integrity check. 

Share this post


Link to post
Share on other sites
  • 0

What version of StableBit CloudDrive are you using?

 

 

 

 

Is read striping enabled? 

 

Also, are you using the UI to confirm access? Or using something like RESMON? 

 **********snip*********

 

 

 

By crash, you mean BSOD? If so, please grab the memory dumps.

http://wiki.covecube.com/StableBit_CloudDrive_System_Crashes

 

And if you're not already, could you see if the 1.0.0.592 works better? 

http://dl.covecube.com/CloudDriveWindows/beta/download/StableBit.CloudDrive_1.0.0.592_x64_BETA.exe

 

 

Ah, you're using an older version, are you not? 

1.0.0.463, correct? 

 

If so, this is due to a weird API issue (meaning, it's one that Google didn't actually include in their well written documentation, and no, there is no sarcasm here, actually).  Specifically, we cannot query the same drive more than 20 times in a short period of time (seconds, IIRC).  It's a known issue and it's been addressed in the newer beta builds. 

 

However, it may require you to recreate the drive, as it's partially an issue with the "format" that the raw data is using. 

 

 

As for Amazon Cloud Drive, again, try the newer build. There are a lot of stability, reliability and performance fixes.  It may cause the Amazon Cloud Drive provider to run better. 

 

And as for re-enabling it, with the newer version, you should be safer to do so. Though, I'd recommend making sure that you use different drives for the cache, to optimize performance. 

 

 

For Google Drive, it may not be necessary.  Some providers, like Amazon Cloud Drive, it's essentially required (it loses data if you don't verify, but if you always verify, it seems to be fine ....)  

 

Basically, it changes the upload process. Instead of clearing out the update data, it downloads it, and verifies that it uploaded properly.  If it did, it clears the data. But if it doesn't, it will reupload it. 

 

If you have the bandwidth (500/500, mbps, I'm assuming), then it's worth having on, as it's another integrity check. 

 

  • Read striping is enabled in the DrivePool and I've checked the DrivePool and Resource Monitor GUI's to check for activity.
  • Yes, BSODS - I tried to grab the memory dumps but WinCrashReport didn't show any...

 

Versions I'm using: (Started using StableBit with those... 3rd day now - and created the drives with them.)

  • CloudDrive: Version 1.0.0.463 BETA
  • DrivePool: Version 2.2.651
  • Scanner: Version 2.5.2.3103 BETA

I thought I downloaded the latest beta's from the web ;S.

 

 

Amazon Cloud Drive: I'll try and re-enable it after I've converted my pools to StableBit... server been down for days now.. and still 10TB to transfer... which with 256KB/s....is hopeless.

 

Another Issue: At times after closing the DrivePool.UI.exe it will not show on re-launch, but does open another process (task manager)

 

As for Chunk Size: Any recommended links to find the proper 'chunk' size for Google Drive/amazon/providers? a search on the forum gave nothing :(

Share this post


Link to post
Share on other sites
  • 0

On the forum everyone is talking about Deduplication on the DrivePool and it may or may not work - but why is no-one using deduplication on the individual drives that form the pool? That way we keep the file duplication across drives, but dedupe on each drive.

e.g. Save 10x *.vhdx to 3 disks (file duplication). And as they are all Ubuntu 16.04, deduplication saves space on each individual volume - while keeping the files on 3 disks as well.

Or did I understand wrong xD

Share this post


Link to post
Share on other sites
  • 0
  • Read striping is enabled in the DrivePool and I've checked the DrivePool and Resource Monitor GUI's to check for activity.
  • Yes, BSODS - I tried to grab the memory dumps but WinCrashReport didn't show any...

 

For the BSOD, check to see if the memory.dmp file exists, (ignore wincrashreport).  Also, try seeing if any minidumps have been created. 

 

As for the read striping, you're definitely not seeing activity in Resource Monitor? 

 

Amazon Cloud Drive: I'll try and re-enable it after I've converted my pools to StableBit... server been down for days now.. and still 10TB to transfer... which with 256KB/s....is hopeless.

 10TB over 256KB/s .... ouch. :(

 

 

Another Issue: At times after closing the DrivePool.UI.exe it will not show on re-launch, but does open another process (task manager)

 

When this happens, could you do this:

 http://wiki.covecube.com/StableBit_DrivePool_UI_Memory_Dump

 

As for Chunk Size: Any recommended links to find the proper 'chunk' size for Google Drive/amazon/providers? a search on the forum gave nothing :(

 

This depends on your bandwidth (upstream and downstream). 

 

Try setting the storage chunk size and the minimum download size to larger values.  This may improve performance, as this may help get up to "full speed" with the download/upload. 

 

http://community.covecube.com/index.php?/topic/1991-clouddrive-googledrive/&do=findComment&comment=13792

 

 

 

On the forum everyone is talking about Deduplication on the DrivePool and it may or may not work - but why is no-one using deduplication on the individual drives that form the pool? That way we keep the file duplication across drives, but dedupe on each drive.

 

e.g. Save 10x *.vhdx to 3 disks (file duplication). And as they are all Ubuntu 16.04, deduplication saves space on each individual volume - while keeping the files on 3 disks as well.

 

Or did I understand wrong xD

 

 

Deduplication on the pool (the large drive itself) would be best. 

The reason for this is that it would get all of the files, and not just some of them.

 

However, because it's not a block based solution, it doesn't work on the pool at all.

 

 

As for the individual drives in the pool, yes it does work. 

And it only works per drive (well, per volume). So if some of those linux VHDs were on a different drive, they wouldn't get dedup-ed "fully".  It would occur per disk. 

The issue here is that StableBit DrivePool's default placement strategy is to place new files on the disk with the most available free space. So the linux VMs may not end up on the same disk, and then couldn't be deduplicated. 

 

That means that you won't save as much space with deduplication, as you would if you were able to dedup the entire pool. 

 

 

But as I said, it does work on the underlying disks. However, there are caveats here, because of how the software works.

 

Since you're using the public beta build, you're fine.  We introduced a check into it that looks for the deduplication filter.  

 

Why is this important?  Deduplication works by *actually removing* the duplicate data in each file and placing the raw, duplicate blocks in the System Volume Info folder. 

It then uses the aforementioned "deduplication filter" to splice the partial file and the rest of the raw data back together, into a single, full file.

 

The issue here is that by default, the pool bypasses the filter for the underlying disks. Meaning that we would only see the unique data, and not the duplicated data in the file.   This is done for performance and compatibility reasons, but it completely breaks deduplication. 

 

Disabling the "bypass file system filters" features can cause issues (and a slight performance hit), it fixes the issue. And the 2.2.0.651 version automatically does this when it detects the Deduplication feature. 

Share this post


Link to post
Share on other sites
  • 0

 

For the BSOD, check to see if the memory.dmp file exists, (ignore wincrashreport).  Also, try seeing if any minidumps have been created. 

 

As for the read striping, you're definitely not seeing activity in Resource Monitor? 

 

 10TB over 256KB/s .... ouch. :(

 

 

 

When this happens, could you do this:

 http://wiki.covecube.com/StableBit_DrivePool_UI_Memory_Dump

 

 

This depends on your bandwidth (upstream and downstream). 

 

Try setting the storage chunk size and the minimum download size to larger values.  This may improve performance, as this may help get up to "full speed" with the download/upload. 

 

http://community.covecube.com/index.php?/topic/1991-clouddrive-googledrive/&do=findComment&comment=13792

 

 

 

 

 

Deduplication on the pool (the large drive itself) would be best. 

The reason for this is that it would get all of the files, and not just some of them.

 

However, because it's not a block based solution, it doesn't work on the pool at all.

 

 

As for the individual drives in the pool, yes it does work. 

And it only works per drive (well, per volume). So if some of those linux VHDs were on a different drive, they wouldn't get dedup-ed "fully".  It would occur per disk. 

The issue here is that StableBit DrivePool's default placement strategy is to place new files on the disk with the most available free space. So the linux VMs may not end up on the same disk, and then couldn't be deduplicated. 

 

That means that you won't save as much space with deduplication, as you would if you were able to dedup the entire pool. 

 

 

But as I said, it does work on the underlying disks. However, there are caveats here, because of how the software works.

 

Since you're using the public beta build, you're fine.  We introduced a check into it that looks for the deduplication filter.  

 

Why is this important?  Deduplication works by *actually removing* the duplicate data in each file and placing the raw, duplicate blocks in the System Volume Info folder. 

It then uses the aforementioned "deduplication filter" to splice the partial file and the rest of the raw data back together, into a single, full file.

 

The issue here is that by default, the pool bypasses the filter for the underlying disks. Meaning that we would only see the unique data, and not the duplicated data in the file.   This is done for performance and compatibility reasons, but it completely breaks deduplication. 

 

Disabling the "bypass file system filters" features can cause issues (and a slight performance hit), it fixes the issue. And the 2.2.0.651 version automatically does this when it detects the Deduplication feature. 

 

By " Disabling the "bypass file system filters" features can cause issues (and a slight performance hit), it fixes the issue. And the 2.2.0.651 version automatically does this when it detects the Deduplication feature." u mean that it will automatically disable/enable the Bypass file system filter option OR regardless of the setting, it will detect the deduplication filter? Right now i have dedup enabled on all individual drives, but 'Bypass File System Filters'  is still enabled.

 

@ BSODS I had both the memory.dmp and minidump files, uploaded both to: http://wiki.covecube.com/StableBit_DrivePool_System_Crashes#Submit_a_crash_dumpwidget.

Share this post


Link to post
Share on other sites
  • 0

The 2.2.0.651 version should disable the "Bypass file system filters" option if the deduplication feature is installed. 

 

The option is under Pool Options -> Performance.  The "Bypass file system filters" option should be unchecked.  if it isn't, then please do uncheck it. 

 

 

 

 

As for the BSODing, this looks to be a hardware related issue, and not strictly related to our software. 

Specifically, some of the crash dumps indicate a DPC Watchdog error, which definitely indicates a hardware issue.

 

I'd recommend running a memory test on the system, to start off with. 

And I'd recommend running a CHKDSK pass of the system disk, to make sure. 

Share this post


Link to post
Share on other sites
  • 0

The 2.2.0.651 version should disable the "Bypass file system filters" option if the deduplication feature is installed. 

 

The option is under Pool Options -> Performance.  The "Bypass file system filters" option should be unchecked.  if it isn't, then please do uncheck it. 

 

 

 

 

As for the BSODing, this looks to be a hardware related issue, and not strictly related to our software. 

Specifically, some of the crash dumps indicate a DPC Watchdog error, which definitely indicates a hardware issue.

 

I'd recommend running a memory test on the system, to start off with. 

And I'd recommend running a CHKDSK pass of the system disk, to make sure. 

Ah okay, did the following:

 

  • CHKDSK on C:/ , seems to have removed the StableBit Scanner notification of filesystem issue
  • DPC Watchdog was probably related to a supermicro driver ( no official support for 2016 yet <i know, shrug>)
  • Manually disabled File System Bypassing....
  • Hyper-V drive checkpoints are now forced to another location ( REFS + SSDs) as they bugged out on the pool
  • Enabled deduplication per drive ( see 100-200GB save atm)
  • Can probably force Hyper-V files 3x duplicated to 3 specific drives & to gain some more dedup savings;
  •  

 

Absolutely liking the flexibility with Folder specific deduplication vs what I achieved on Storage Spaces

 

  • Hyper-V VM's are still a bit slow though;
  • Some won't start as an individual drive is full - yet pool shows space (prob have to wait for balancing);

@ CloudDrive:

 

  • All Cloud Drives have a dedicated 64GB SSD for local cache with 10GB pinned and 'expandable' as config, formatted as NTFS. ( Tried NTFS + dedup.. it showed an odd 2TB savings...xD - disabled it)
  • I'll probably pick Folder specific duplication & x drives, to have 2 copies go to two seperate clouds, and 2 local
  • OR i'll have another tool / sync from the DrivePool to the Cloud;
  • Amazon keeps detaching itself... i'll try later on.

 

@Scanner

  • I have enabled some of the heat throttling options ( works great as my cache drives shoot up to 60*C on bulk transfers);
  • It keeps throttling with 'bus saturated', but an M1015 SAS with only 8 HDD's (WD Reds)... doing 120MB/s... don't think that saturates it? Any way we can manually configure that bus limit?
  • EDIT: Found your recommendations on: http://community.covecube.com/index.php?/topic/1990-difference-between-versions/?hl=backupand disabled 'do not interfere with same bus' and scans turbo'd up from 120 to 450MB/s per disk ( /B/sec DISK I/O

Thanks for the great responses!

Share this post


Link to post
Share on other sites

Create an account or sign in to comment

You need to be a member in order to leave a comment

Create an account

Sign up for a new account in our community. It's easy!

Register a new account

Sign in

Already have an account? Sign in here.

Sign In Now


×