-
Notifications
You must be signed in to change notification settings - Fork 634
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Feature]: Datanode support to manage both SSD and HDD disks #3239
Comments
We have a rough idea :
|
we have roughly implement a draft version , and verify the idea. #3243 The local transition process described above is faster than the original and even more faster than file upload process. In my environment(4c,8G ,1000M netwrok),a 4GB file local migration takes about more than 10 seconds(even faster than s3 put which takes about 30 seconds) , and the original cross-node migration process takes about 90 seconds. |
This solution requires significant changes, and we can only consider whether to incorporate this approach after the completion of the first phase of the HybridCloud project. |
Contact Details
[email protected]
Is there an existing issue for this?
Is your feature request related to a problem? Please describe.
In our use cases, there are only few machines available to deploy cfs cluster , eg 3 or 5. Those machines are both equipped with SSD and HDD disks. For some cold data, we want to migrate them from SSD to HDD.
As I know, the developing Hybrid-cloud branch support data cool-down feature, but need to deploy at least two zones , in each zone the nodes are limited to configure same type of disks. when migrating, the cold data are transfered from SSD zone to HDD zone.
The solution now is not friendly for small cfs clusters, eg. nodes less than 6 . Also, for machines with both SSD and HDD disks, the migration process can not leverage Locality to imporve performance.
Describe the solution you'd like.
As a solutiion, I think the datanode should have the ability to manage both SSD and HDD which seems more reasonable. As a result, even the cluster with 3 nodes (with both SSD and HDD, one zone) can experience the data cool-down feature, Meanwhile, cubefs can optimize cold data migration performance. When migrating cold data, prefer to choose HDD directories that on the same node as the destination, which can reduce network traffic and improve migration performance greatly.
Describe an alternate solution.
As a alternate solution, we can deploy cfs in containers, so we can deploy two zones on three nodes cluster. Nodes in each zone manage a single type of disks. Nodes configure secondary IP also can help.
But can't use locality to reduce network traffics caused by data migration
Anything else? (Additional Context)
No response
The text was updated successfully, but these errors were encountered: