Movatterモバイル変換


[0]ホーム

URL:


Jump to content
WikipediaThe Free Encyclopedia
Search

Ceph (software)

From Wikipedia, the free encyclopedia
Open-source storage platform
This article is about the computer storage platform. For other uses, seeCeph (disambiguation).
Ceph Storage
Original author(s)Inktank Storage:
  • Sage Weil
  • Yehuda Sadeh Weinraub
  • Gregory Farnum
  • Josh Durgin
  • Samuel Just
  • Wido den Hollander
Developer(s)
Stable release
19.2.2[2] Edit this on Wikidata (Squid) / 10 April 2025
Repository
Written inC++,Python[3]
Operating systemLinux,FreeBSD,[4]Windows[5]
TypeDistributed object store
LicenseLGPLv2.1[6]
Websiteceph.io

Ceph (pronounced/ˈsɛf/) is afree andopen-source software-definedstorageplatform that providesobject storage,[7]block storage, andfile storage built on a commondistributed cluster foundation. Ceph provides distributed operation without asingle point of failure and scalability to theexabyte level. Since version 12 (Luminous), Ceph does not rely on any other conventional filesystem and directly managesHDDs andSSDs with its own storage backend BlueStore and can expose aPOSIXfilesystem.

Cephreplicates data withfault tolerance,[8] usingcommodity hardware and Ethernet IP and requiring no specific hardware support. Ceph ishighly available and ensures strong data durability through techniques including replication,erasure coding, snapshots and clones. By design, the system is both self-healing andself-managing, minimizing administration time and other costs.

Large-scale production Ceph deployments includeCERN,[9][10]OVH[11][12][13][14] andDigitalOcean.[15][16]

Design

[edit]
A high-level overview of the Ceph's internal organization[17]: 4 

Ceph employs five distinct kinds ofdaemons:[17]

  • Cluster monitors (ceph-mon) that keep track of active and failed cluster nodes, cluster configuration, and information about data placement and global cluster state.
  • OSDs (ceph-osd) that manage bulk data storage devices directly via the BlueStore back end,[18] which since the v12.x release replaces the Filestore[19] back end, which was implemented on top of a traditional filesystem)
  • Metadata servers (ceph-mds) that maintain and broker access toinodes anddirectories inside a CephFS filesystem
  • HTTP gateways (ceph-rgw) that expose the object storage layer as an interface compatible withAmazon S3 orOpenStack Swift APIs
  • Managers (ceph-mgr) that perform cluster monitoring, bookkeeping, and maintenance tasks, and interface to external monitoring systems and management (e.g. balancer, dashboard,Prometheus, Zabbix plugin)[20]

All of these are fully distributed, and may be deployed on disjoint, dedicated servers or in aconverged topology. Clients with different needs directly interact with appropriate cluster components.[21]

Cephdistributes data across multiple storage devices and nodes to achieve higher throughput, in a fashion similar toRAID. Adaptiveload balancing is supported whereby frequently accessed services may be replicated over more nodes.[22]

As of September 2017[update], BlueStore is the default and recommended storage back end for production environments,[23] which provides better latency and configurability than the older Filestore back end, and avoiding the shortcomings of filesystem based storage involving additional processing and caching layers. The Filestore back end will be deprecated as of the Reef release in mid 2023.XFS was the recommended underlying filesystem for Filestore OSDs, andBtrfs could be used at one's own risk.ext4 filesystems were not recommended due to limited metadata capacity.[24] The BlueStore back end does still use XFS for a small metadata partition.[25]

Object storage S3

[edit]
An architecture diagram showing the relations among components of the Ceph storage platform

Ceph implements distributedobject storage via the RADOS GateWay (ceph-rgw), which exposes the underlying storage layer via an interface compatible withAmazon S3 orOpenStack Swift.

Ceph RGW deployments scale readily and often utilize large and dense storage media for bulk use cases that includeBig Data (datalake),backups &archives,IOT, media, video recording, and deployment images forvirtual machines andcontainers.[26]

Ceph's software libraries provide client applications with direct access to thereliable autonomic distributed object store (RADOS) object-based storage system. More frequently used are libraries for Ceph'sRADOS Block Device (RBD),RADOS Gateway, andCeph File System services. In this way, administrators can maintain their storage devices within a unified system, which makes it easier to replicate and protect the data.

The "librados"software libraries provide access inC,C++,Java,PHP, andPython. The RADOS Gateway also exposes the object store as aRESTful interface which can present as both nativeAmazon S3 andOpenStack Swift APIs.

Block storage

[edit]

Ceph can provide clients withthin-provisionedblock devices. When an application writes data to Ceph using a block device, Ceph automatically stripes and replicates the data across the cluster. Ceph'sRADOS Block Device (RBD) also integrates withKernel-based Virtual Machines (KVMs).

Ceph block storage may be deployed on traditional HDDs and/orSSDs which are associated with Ceph's block storage for use cases, including databases, virtual machines, data analytics, artificial intelligence, and machine learning. Block storage clients often require highthroughput andIOPS, thus Ceph RBD deployments increasingly utilize SSDs withNVMe interfaces.

"RBD" is built on with Ceph's foundational RADOS object storage system that provides the librados interface and the CephFS file system. Since RBD is built on librados, RBD inherits librados's abilities, including clones andsnapshots. By striping volumes across the cluster, Ceph improves performance for large block device images.

"Ceph-iSCSI" is a gateway which enables access to distributed, highly available block storage fromMicrosoft Windows andVMware vSphere servers or clients capable of speaking theiSCSI protocol. By using ceph-iscsi on one or more iSCSI gateway hosts, Ceph RBD images become available as Logical Units (LUs) associated with iSCSI targets, which can be accessed in an optionally load-balanced, highly available fashion.

Since ceph-iscsi configuration is stored in the Ceph RADOS object store, ceph-iscsi gateway hosts are inherently without persistent state and thus can be replaced, augmented, or reduced at will. As a result, Ceph Storage enables customers to run a truly distributed, highly-available, resilient, and self-healing enterprise storage technology on commodity hardware and an entirely open source platform.

The block device can be virtualized, providing block storage to virtual machines, in virtualization platforms such asOpenShift,OpenStack,Kubernetes,OpenNebula,Ganeti,Apache CloudStack andProxmox Virtual Environment.

File storage

[edit]

Ceph's file system (CephFS) runs on top of the same RADOS foundation as Ceph's object storage and block device services. The CephFS metadata server (MDS) provides a service that maps the directories and file names of the file system to objects stored within RADOS clusters. The metadata server cluster can expand or contract, and it can rebalance file system metadata ranks dynamically to distribute data evenly among cluster hosts. This ensures high performance and prevents heavy loads on specific hosts within the cluster.

Clients mount thePOSIX-compatible file system using aLinux kernel client. An olderFUSE-based client is also available. The servers run as regular Unixdaemons.

Ceph's file storage is often associated with log collection, messaging, and file storage.

Dashboard

[edit]
Ceph Dashboard landing page (2023)

From 2018 there is also a Dashboard web UI project, which helps to manage the cluster. It's being developed by Ceph community on LGPL-3 and usesceph-mgr,Python,Angular andGrafana.[27] Its landing page has been refreshed in the beginning of 2023.[28]

Previous dashboards were developed but are closed now: Calamari (2013–2018), OpenAttic (2013–2019), VSM (2014–2016), Inkscope (2015–2016) and Ceph-Dash (2015–2017).[29]

Crimson

[edit]

Beginning in 2019 the Crimson project has been reimplementing the OSD data path. The goal of Crimson is to minimize latency and CPU overhead. Modern storage devices and interfaces includingNVMe and3D XPoint have become much faster thanHDD and even SAS/SATASSDs, but CPU performance has not kept pace. Moreovercrimson-osd is meant to be a backward-compatibledrop-in replacement forceph-osd. While Crimson can work with the BlueStore back end (via AlienStore), a new native ObjectStore implementation called SeaStore is also being developed along with CyanStore for testing purposes. One reason for creating SeaStore is that transaction support in the BlueStore back end is provided byRocksDB, which needs to be re-implemented to achieve better parallelism.[30][31][32]

History

[edit]

Ceph was created bySage Weil for hisdoctoral dissertation,[33] which was advised by Professor Scott A. Brandt at theJack Baskin School of Engineering,University of California, Santa Cruz (UCSC), and sponsored by theAdvanced Simulation and Computing Program (ASC), includingLos Alamos National Laboratory (LANL),Sandia National Laboratories (SNL), andLawrence Livermore National Laboratory (LLNL).[34] The first line of code that ended up being part of Ceph was written by Sage Weil in 2004 while at a summer internship at LLNL, working on scalable filesystem metadata management (known today as Ceph's MDS).[35] In 2005, as part of a summer project initiated by Scott A. Brandt and led by Carlos Maltzahn, Sage Weil created a fully functional file system prototype which adopted the name Ceph. Ceph made its debut with Sage Weil giving two presentations in November 2006, one atUSENIX OSDI 2006[36] and another atSC'06.[37]

After his graduation in autumn 2007, Weil continued to work on Ceph full-time, and the core development team expanded to include Yehuda Sadeh Weinraub and Gregory Farnum. On March 19, 2010,Linus Torvalds merged the Ceph client into Linux kernel version 2.6.34[38][39] which was released on May 16, 2010. In 2012, Weil createdInktank Storage for professional services and support for Ceph.[40][41]

In April 2014,Red Hat purchased Inktank, bringing the majority of Ceph development in-house to make it a production version for enterprises with support (hotline) and continuous maintenance (new versions).[42]

In October 2015, the Ceph Community Advisory Board was formed to assist the community in driving the direction of open source software-defined storage technology. The charter advisory board includes Ceph community members from global IT organizations that are committed to the Ceph project, including individuals fromRed Hat,Intel,Canonical,CERN,Cisco,Fujitsu,SanDisk, andSUSE.[43]

In November 2018, the Linux Foundation launched the Ceph Foundation as a successor to the Ceph Community Advisory Board. Founding members of the Ceph Foundation included Amihan,Canonical,China Mobile,DigitalOcean,Intel,OVH, ProphetStor Data Services,Red Hat, SoftIron,SUSE,Western Digital, XSKY Data Technology, andZTE.[44]

In March 2021, SUSE discontinued its Enterprise Storage product incorporating Ceph in favor ofRancher's Longhorn,[45] and the former Enterprise Storage website was updated stating "SUSE has refocused the storage efforts around serving our strategic SUSE Enterprise Storage Customers and are no longer actively selling SUSE Enterprise Storage."[46]

Release history

[edit]
Release history
NameReleaseFirst releaseEnd of
life
Milestones
ArgonautUnsupported: 0.48July 3, 2012First major "stable" release
BobtailUnsupported: 0.56January 1, 2013
CuttlefishUnsupported: 0.61May 7, 2013ceph-deploy is stable
DumplingUnsupported: 0.67August 14, 2013May 2015namespace, region, monitoring REST API
EmperorUnsupported: 0.72November 9, 2013May 2014multi-datacenter replication for RGW
FireflyUnsupported: 0.80May 7, 2014April 2016erasure coding, cache tiering, primary affinity, key/value OSD backend (experimental), standalone RGW (experimental)
GiantUnsupported: 0.87October 29, 2014April 2015
HammerUnsupported: 0.94April 7, 2015August 2017
InfernalisUnsupported: 9.2.0November 6, 2015April 2016
JewelUnsupported: 10.2.0April 21, 20162018-06-01Stable CephFS, experimental OSD back end named BlueStore, daemons no longer run as the root user
KrakenUnsupported: 11.2.0January 20, 20172017-08-01BlueStore is stable, EC for RBD pools
LuminousUnsupported: 12.2.0August 29, 20172020-03-01pg-upmap balancer
MimicUnsupported: 13.2.0June 1, 20182020-07-22snapshots are stable, Beast is stable, official GUI (Dashboard)
NautilusUnsupported: 14.2.0March 19, 20192021-06-01asynchronous replication, auto-retry of failed writes due to grown defect remapping
OctopusUnsupported: 15.2.0March 23, 20202022-06-01
PacificUnsupported: 16.2.0March 31, 2021[47]2023-06-01
QuincyUnsupported: 17.2.0April 19, 2022[48]2024-06-01auto-setting of min_alloc_size for novel media
ReefSupported: 18.2.0Aug 3, 2023[49]2025-08-01[50]
SquidLatest version:19.2.0Sep 26, 2024[51]2026-09-19[52]
Tentacle[53]Future version: TBATBA
Legend:
Unsupported
Supported
Latest version
Preview version
Future version

Available platforms

[edit]

While basically built forLinux, Ceph has been also partially ported to Windows platform. It is production-ready forWindows Server 2016 (some commands might be unavailable due to lack ofUNIX socket implementation),Windows Server 2019 andWindows Server 2022, but testing/development can be done also onWindows 10 andWindows 11. One can use Ceph RBD and CephFS on Windows, but OSD is not supported on this platform.[54][5][55]

There is alsoFreeBSD implementation of Ceph.[4]

MicroCeph

[edit]

MicroCeph is a simplified Ceph deployment system for non-experts usingsnap packaging system, created byCanonical in 2022.[56] It's isolated from the underlying host, platform independent, scalable and offers minimal setup and maintenance overheads.[57] MicroCeph supports all Ceph data access protocols - block, file and object - and can be deployed with full disk encryption.[58]

Etymology

[edit]

The name "Ceph" is a shortened form of "cephalopod", a class ofmolluscs that includes squids, cuttlefish, nautiloids, and octopuses. The name (emphasized by the logo) suggests the highly parallel behavior of an octopus and was chosen to associate the file system with "Sammy", thebanana slug mascot ofUCSC.[17] Both cephalopods and banana slugs are molluscs.

See also

[edit]

References

[edit]
  1. ^"Ceph Community Forms Advisory Board". 2015-10-28. Archived fromthe original on 2019-01-29. Retrieved2016-01-20.
  2. ^https://ceph.io/en/news/blog/2025/v19-2-2-squid-released/. Retrieved29 June 2025.{{cite web}}:Missing or empty|title= (help)
  3. ^"GitHub Repository".GitHub.Archived from the original on 2023-05-26. Retrieved2017-02-08.
  4. ^ab"FreeBSD Quarterly Status Report".Archived from the original on 2020-12-23. Retrieved2017-02-27.
  5. ^ab"Installing Ceph on Windows".Ceph.Archived from the original on 2 July 2023. Retrieved2 July 2023.
  6. ^"LGPL2.1 license file in the Ceph sources".GitHub. 2014-10-24.Archived from the original on 2020-12-09. Retrieved2014-10-24.
  7. ^Nicolas, Philippe (2016-07-15)."The History Boys: Object storage ... from the beginning".The Register.Archived from the original on 2020-03-14. Retrieved2021-01-25.
  8. ^Jeremy Andrews (2007-11-15)."Ceph Distributed Network File System".KernelTrap. Archived fromthe original on 2007-11-17. Retrieved2007-11-15.
  9. ^"Ceph Clusters".CERN. Archived fromthe original on 12 November 2022. Retrieved12 November 2022.
  10. ^"Ceph Operations at CERN: Where Do We Go From Here? - Dan van der Ster & Teo Mouratidis, CERN".YouTube. 24 May 2019.Archived from the original on 12 November 2022. Retrieved12 November 2022.
  11. ^Dorosz, Filip (15 June 2020)."Journey to next-gen Ceph storage at OVHcloud with LXD".OVHcloud.Archived from the original on 12 November 2022. Retrieved12 November 2022.
  12. ^"CephFS distributed filesystem".OVHcloud. Retrieved12 November 2022.
  13. ^"Ceph - Distributed Storage System in OVH [en] - Bartłomiej Święcki".YouTube. 7 April 2016.Archived from the original on 12 November 2022. Retrieved12 November 2022.
  14. ^"200 Clusters vs 1 Admin - Bartosz Rabiega, OVH".YouTube. 24 May 2019.Archived from the original on 15 November 2022. Retrieved15 November 2022.
  15. ^D'Atri, Anthony (31 May 2018)."Why We Chose Ceph to Build Block Storage".DigitalOcean. Retrieved12 November 2022.
  16. ^"Ceph Tech Talk: Ceph at DigitalOcean".YouTube. 7 October 2021.Archived from the original on 12 November 2022. Retrieved12 November 2022.
  17. ^abcM. Tim Jones (2010-06-04)."Ceph: A Linux petabyte-scale distributed file system"(PDF).IBM.Archived(PDF) from the original on 2016-03-04. Retrieved2014-12-03.
  18. ^"BlueStore". Ceph.Archived from the original on 2019-01-28. Retrieved2017-09-29.
  19. ^"BlueStore Migration". Archived fromthe original on 2019-12-04. Retrieved2020-04-12.
  20. ^"Ceph Manager Daemon — Ceph Documentation".docs.ceph.com. Archived fromthe original on June 6, 2018. Retrieved2019-01-31.archive linkArchived June 19, 2020, at theWayback Machine
  21. ^Jake Edge (2007-11-14)."The Ceph filesystem".LWN.net.Archived from the original on 2023-06-23. Retrieved2016-07-21.
  22. ^Anthony D'Atri, Vaibhav Bhembre (2017-10-01)."Learning Ceph, Second Edition".Packt.Archived from the original on 2023-05-09. Retrieved2023-05-09.
  23. ^Sage Weil (2017-08-29)."v12.2.0 Luminous Released". Ceph Blog.Archived from the original on 2023-04-17. Retrieved2017-09-01.
  24. ^"Hard Disk and File System Recommendations". ceph.com. Archived fromthe original on 2017-07-14. Retrieved2017-06-26.
  25. ^"BlueStore Config Reference". Archived fromthe original on July 20, 2019. RetrievedApril 12, 2020.
  26. ^"10th International Conference "Distributed Computing and Grid Technologies in Science and Education" (GRID'2023)".JINR (Indico). 2023-07-03. Retrieved2023-08-09.
  27. ^"Ceph Dashboard".Ceph documentation.Archived from the original on 11 April 2023. Retrieved11 April 2023.
  28. ^Gomez, Pedro Gonzalez (23 February 2023)."Introducing the new Dashboard Landing Page". Retrieved11 April 2023.
  29. ^"Operating Ceph from the Ceph Dashboard: past, present and future".YouTube. 22 November 2022.Archived from the original on 11 April 2023. Retrieved11 April 2023.
  30. ^Just, Sam (18 January 2021)."Crimson: evolving Ceph for high performance NVMe".Red Hat Emerging Technologies. Retrieved12 November 2022.
  31. ^Just, Samuel (10 November 2022)."What's new with Crimson and Seastore?".YouTube.Archived from the original on 12 November 2022. Retrieved12 November 2022.
  32. ^"Crimson: Next-generation Ceph OSD for Multi-core Scalability".Ceph blog. Ceph. 7 February 2023.Archived from the original on 30 March 2023. Retrieved11 April 2023.
  33. ^Sage Weil (2007-12-01)."Ceph: Reliable, Scalable, and High-Performance Distributed Storage"(PDF).University of California, Santa Cruz. Archived fromthe original(PDF) on 2017-07-06. Retrieved2017-03-11.
  34. ^Gary Grider (2004-05-01)."The ASCI/DOD Scalable I/O History and Strategy"(PDF).University of Minnesota. Retrieved2019-07-17.
  35. ^Dynamic Metadata Management for Petabyte-Scale File Systems, SA Weil, KT Pollack, SA Brandt, EL Miller, Proc. SC'04, Pittsburgh, PA, November, 2004
  36. ^"Ceph: A scalable, high-performance distributed file system," SA Weil, SA Brandt, EL Miller, DDE Long, C Maltzahn, Proc. OSDI, Seattle, WA, November, 2006
  37. ^"CRUSH: Controlled, scalable, decentralized placement of replicated data," SA Weil, SA Brandt, EL Miller, DDE Long, C Maltzahn, SC'06, Tampa, FL, November, 2006
  38. ^Sage Weil (2010-02-19)."Client merged for 2.6.34". ceph.newdream.net. Archived fromthe original on 2010-03-23. Retrieved2010-03-21.
  39. ^Tim Stephens (2010-05-20)."New version of Linux OS includes Ceph file system developed at UCSC". news.ucsc.edu.
  40. ^Bryan Bogensberger (2012-05-03)."And It All Comes Together". Inktank Blog. Archived fromthe original on 2012-07-19. Retrieved2012-07-10.
  41. ^Joseph F. Kovar (July 10, 2012)."The 10 Coolest Storage Startups Of 2012 (So Far)".CRN.Archived from the original on November 5, 2013. RetrievedJuly 19, 2013.
  42. ^Red Hat Inc (2014-04-30)."Red Hat to Acquire Inktank, Provider of Ceph". Red Hat.Archived from the original on 2016-02-01. Retrieved2014-08-19.
  43. ^"Ceph Community Forms Advisory Board". 2015-10-28. Archived fromthe original on 2019-01-29. Retrieved2016-01-20.
  44. ^"The Linux Foundation Launches Ceph Foundation To Advance Open Source Storage". 2018-11-12.[permanent dead link]
  45. ^"SUSE says tschüss to Ceph-based enterprise storage product – it's Rancher's Longhorn from here on out".Archived from the original on 2021-04-21. Retrieved2021-05-29.
  46. ^"SUSE Enterprise Software-Defined Storage".Archived from the original on 2020-09-27. Retrieved2021-05-29.
  47. ^Ceph.io — v16.2.0 Pacific released
  48. ^Ceph.io — v17.2.0 Quincy released
  49. ^Flores, Laura (6 August 2023)."v18.2.0 Reef released".Ceph Blog.Archived from the original on 26 August 2023. Retrieved26 August 2023.
  50. ^"Active Releases".docs.ceph.com. Retrieved3 October 2024.
  51. ^Flores, Laura (26 September 2024)."v19.2.0 Squid released".Ceph Blog.Archived from the original on 27 September 2024. Retrieved3 October 2024.
  52. ^"Active Releases".docs.ceph.com. Retrieved3 October 2024.
  53. ^"Ceph Developer Summit - Tentacle".ceph.io. 11 August 2024.Archived from the original on 17 September 2024. Retrieved3 October 2024.
  54. ^"Ceph for Windows". Cloudbase Solutions. Retrieved2 July 2023.
  55. ^Pilotti, Alessandro (4 May 2023)."Ceph on Windows".YouTube. Retrieved2 July 2023.
  56. ^Graber, Stéphane."Introducing MicroCeph".Linux Containers Forum.Archived from the original on 17 December 2024. Retrieved24 June 2025.
  57. ^Williams, Phil (June 2025)."What's new in MicroCeph?"(PDF).ceph.io. Retrieved24 June 2025.
  58. ^Williams, Philip."Here, there, everywhere – MicroCeph makes edge storage easy".ubuntu.com. Canonical Ltd. Retrieved24 June 2025.

Further reading

[edit]

External links

[edit]
Look upκεφαλή in Wiktionary, the free dictionary.
Wikimedia Commons has media related toCeph.
Disk and
non-rotating
Optical disc
Flash memory andSSD
host-sidewear leveling
Distributed parallel
NAS
Specialized
Pseudo
Encrypted
Types
Features
Access control
Interfaces
Lists
Layouts
Major products
Services
Projects
Defunct
People
Mergers and acquisitions
Related
Retrieved from "https://en.wikipedia.org/w/index.php?title=Ceph_(software)&oldid=1297493946"
Categories:
Hidden categories:

[8]ページ先頭

©2009-2025 Movatter.jp