Search the Community
Showing results for tags 'millions'.
Found 2 results
Hi all! Our company has websites, which host millions of files. These are 10-50 KB each. One Windows server carries around 100 million jpeg-images with a total of around 1 TB disk space. To have a fail-over (or a round robin), we would like to sync files between Windows servers (WAN, not LAN). HOwever, we saw that BitTorrent sync requires around 300 bytes / file. This would mean a memory requirement of 30 GB for BitTorrent alone. Is there a solution to this? Thanks so much!
Hi, I started using BTSync as soon as it came out publically at home and I love it. I use it to sync my music, documents, etc. between multiple computers at home and remotely so that my main system can act as a backup server for all of the others. Some of the folders are ~300GB in 150,000 files, all Windows machines. Now on to my problem: After loving the product for personal use, I wanted to use it at work for a new project. We just bought 5 new Haswell servers and we wanted their configurations to all be the same. -Each has an SSD with the OS mirrored across all of them and everything with that is fine. -They also each have a 3TB HDD (WD RED) to store our compile tools. -All on the same gigabit switch. -iptables firewall disabled. I wanted to use BTSync to keep a folder (/usr/tools/) the same across all servers. Ideally, if new tools are added they would instantly sync to all the other servers, and this was supposed to cut down on my setup time. I have btsync installed on all systems and can view the webpage to manage them. When I go to add the folder (/usr/tools/) there is significant lag before the website registers that it is complete. I let this system finish indexing all the files (239GB in 2,400,000 files) which took over 8 hours (!!!!). I could not believe that it took 8 hours using 100% of one of the CPUs. When I added the second system to the swarm, it was immediately found and added, but the transfer rate was pathetic. It did about 2GB in an hour. I added the other three systems to see if that would help, but it only made things slower. I ended up copying the folders over manually and letting them all index before connecting them back to the LAN. Four of the five servers now report that they are in sync, but one is fully indexed yet still things it needs all 239GB from the other servers (but doesn't transfer anything). With only the four that are in sync online, btsync is still using 100% of one of the cores (presumably trying to constantly check for changes?) So my questions are: -Is it expected that indexing 2.4 million files will take a very very long time? -Is there anyway to let it use more than one core for indexing? -What could be preventing one of the servers from syncing if they are all configured exactly the same? -Would smaller folders be handled better? -Why is the CPU usage so high?