Jump to content

kts

Members
  • Posts

    4
  • Joined

  • Last visited

Recent Profile Visitors

328 profile views

kts's Achievements

Newbie

Newbie (1/3)

0

Reputation

  1. We're beginning to run into a problem with the measuring process. Apparently it takes too long to finish (takes several days up to a week?!?) and with about 2/3 of the drives measured it starts from scratch. Is this behaviour normal? I resetted the settings already and also increased CoveFs_AsyncCriticalThreads and CoveFs_AsyncNormalThreads. Although done, there is not much activity during measuring. Also got the BackgroundTaskPriority up to 12 meanwhile. Is there a way to increase the time in between measuring and especially speed up the measuring process itself? DP service uses like 0,1% of CPU and the measured HDD is at around 100kb/s.
  2. Just wanted to let you know it's working now. Checking takes several days but everything seems to be doing good so far. Performance is good even with many small files.
  3. Version 2.1.1.561 Windows 8.1 I restarted again, and the process started from the beginning. After reading your reply I restarted the service, it kept the statistics. After changing the priority and restarting the service it lost them again, and is now indexing quickly. Strangely it counts to a 2 or 3 figure number and then drops down to close to zero again (I must say that I'm continuing to add data and assume this is the problem). I'll give it a while to calm down. If it will finish I may restart again to see if that is a permanent problem. Do incomplete statistics impair or slow drivepool down? As far as I can see (and hope) balancing and duplication is still working.
  4. Hi, we're testing drivepool to determine if it might be an suitable solution to keep a second online copy and backup. We're dealing with huge numbers of small files (500kb - 25mb) in several "main" folders each with 4096 subfolders. Now we got quite a few of those main folders with between 10 and 50 million files each and adding 25-100k files per day. Additionally we have to import other systems which may add 10-30 million files at once. We're planning for 24 or 36 drive systems with drivepool and can see the 1 billion file barrier break on single systems. I've set up a test machine (i7, 32gig and 6x4tb 7.2k drives exclusively for storage) and gave it a taste of around 50-60 million files with duplication (2x). It did not to bad until I restarted the system. It is working but the dashboard is calculating for a couple of hours now and claiming xxx GB are not duplicated (while they are, number is increasing, I assume it has to check _every_ file on disk?). Will give it a couple of hours more to get organized. Is if drivepool is suitable for those large number of files? I assume there is some kind of database keeping records where which files are placed. Is there a maximum database size or overall file limit per pool? Do those numbers have a known impact on performance? Does this recalculation occur after every restart? Anyone having experience with similar file numbers? Thanks!
×
×
  • Create New...