Been using UNIX since the late 80s; Linux since the mid-90s; virtualization since the early 2000s and spent the past few years working in the cloud space.
Location
Alexandria, VA, USA
Education
B.S. Psychology from Pennsylvania State University
If you're uploading large data-streams, it's best to override the default cipher to one better-optimized to that use-case (Blowfish used to be a good choice)
If you're needing to transfer a bunch of files in a bunch of directories, you can sometimes (e.g., if you're going over a WAN link that implements session-limits …and annoyingly common configuration problem) get better performance by running rsync in concert with the parallel command
Speaking of performance…
If someone ever complains about network throughput speeds, never use SSH-enabled techniques for bandwidth testing. The encryption-overhead of SSH means that such tests will never really show you your network's actual capabilities (unless your network is so degraded that SSH's encryption is no longer a bottleneck).
With respect to rsync:
rsync
in concert with theparallel
commandSpeaking of performance…
If someone ever complains about network throughput speeds, never use SSH-enabled techniques for bandwidth testing. The encryption-overhead of SSH means that such tests will never really show you your network's actual capabilities (unless your network is so degraded that SSH's encryption is no longer a bottleneck).
These are interesting tips. Thanks.