I don't understand your question? You created a 1G RBD/disk and it's full.
You are able to grow it though - but that's a Linux management issue, not
ceph.

As everything is thin-provisioned you can create a RBD with an arbitrary
size - I've create one with 1PB when the cluster only had 600G/Raw
available.

On Mon, Aug 17, 2015 at 1:18 PM, gjprabu <gjpr...@zohocorp.com> wrote:

> Hi All,
>
>            Anybody can help on this issue.
>
> Regards
> Prabu
>
> ---- On Mon, 17 Aug 2015 12:08:28 +0530 *gjprabu <gjpr...@zohocorp.com
> <gjpr...@zohocorp.com>>* wrote ----
>
> Hi All,
>
>                Also please find osd information.
>
> ceph osd dump | grep 'replicated size'
> pool 2 'repo' replicated size 2 min_size 2 crush_ruleset 0 object_hash
> rjenkins pg_num 126 pgp_num 126 last_change 21573 flags hashpspool
> stripe_width 0
>
> Regards
> Prabu
>
>
>
>
> ---- On Mon, 17 Aug 2015 11:58:55 +0530 *gjprabu <gjpr...@zohocorp.com
> <gjpr...@zohocorp.com>>* wrote ----
>
>
>
> Hi All,
>
>    We need to test three OSD and one image with replica 2(size 1GB). While
> testing data is not writing above 1GB. Is there any option to write on
> third OSD.
>
> *ceph osd pool get  repo  pg_num*
> *pg_num: 126*
>
> *# rbd showmapped *
> *id pool image          snap device    *
> *0  rbd  integdownloads -    /dev/rbd0 *-- *Already one*
> *2  repo integrepotest  -    /dev/rbd2  -- newly created*
>
>
> [root@hm2 repository]# df -Th
> Filesystem           Type      Size  Used Avail Use% Mounted on
> /dev/sda5            ext4      289G   18G  257G   7% /
> devtmpfs             devtmpfs  252G     0  252G   0% /dev
> tmpfs                tmpfs     252G     0  252G   0% /dev/shm
> tmpfs                tmpfs     252G  538M  252G   1% /run
> tmpfs                tmpfs     252G     0  252G   0% /sys/fs/cgroup
> /dev/sda2            ext4      488M  212M  241M  47% /boot
> /dev/sda4            ext4      1.9T   20G  1.8T   2% /var
> /dev/mapper/vg0-zoho ext4      8.6T  1.7T  6.5T  21% /zoho
> /dev/rbd0            ocfs2     977G  101G  877G  11% /zoho/build/downloads
> */dev/rbd2            ocfs2    1000M 1000M     0 100%
> /zoho/build/repository*
>
> @:~$ scp -r sample.txt root@integ-hm2:/zoho/build/repository/
> root@integ-hm2's password:
> sample.txt
> 100% 1024MB   4.5MB/s   03:48
> scp: /zoho/build/repository//sample.txt: *No space left on device*
>
> Regards
> Prabu
>
>
>
>
> ---- On Thu, 13 Aug 2015 19:42:11 +0530 *gjprabu <gjpr...@zohocorp.com
> <gjpr...@zohocorp.com>>* wrote ----
>
>
>
> Dear Team,
>
>          We are using two ceph OSD with replica 2 and it is working
> properly. Here my doubt is (Pool A -image size will be 10GB) and its
> replicated with two OSD, what will happen suppose if the size reached the
> limit, Is there any chance to make the data to continue writing in another
> two OSD's.
>
> Regards
> Prabu
>
>
>
>
>
>
>
> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to