gluster / gstatus Goto Github PK
View Code? Open in Web Editor NEWgstatus is a command to view current health of a glusterfs cluster
License: GNU General Public License v3.0
gstatus is a command to view current health of a glusterfs cluster
License: GNU General Public License v3.0
Hi.
I've just cloned the repo to Ubunu 14.04-based and to Fedora 22-based gluster nodes.
I try to run the script without installation and on both systems I get:
$ ./gstatus.py -s
Traceback (most recent call last):
File "./gstatus.py", line 221, in
main()
File "./gstatus.py", line 132, in main
cluster.initialise()
File "/home/afunix/opt/gstatus/gstatus/libgluster/cluster.py", line 91, in initialise
self.define_nodes()
File "/home/afunix/opt/gstatus/gstatus/libgluster/cluster.py", line 140, in define_nodes
cmd.run()
File "/home/afunix/opt/gstatus/gstatus/libcommand/glustercmd.py", line 100, in run
xmldoc = ETree.fromstring(''.join(self.stdout))
File "/usr/lib64/python2.7/xml/etree/ElementTree.py", line 1301, in XML
return parser.close()
File "/usr/lib64/python2.7/xml/etree/ElementTree.py", line 1654, in close
self._raiseerror(v)
File "/usr/lib64/python2.7/xml/etree/ElementTree.py", line 1506, in _raiseerror
raise err
xml.etree.ElementTree.ParseError: no element found: line 1, column 0
The commit is:
commit 858d2f2
Author: Paul Cuzner [email protected]
Date: Mon May 25 17:12:07 2015 +1200
Added licence file
Ubuntu:
$ cat /etc/lsb-release
DISTRIB_ID=Ubuntu
DISTRIB_RELEASE=14.04
DISTRIB_CODENAME=trusty
DISTRIB_DESCRIPTION="Ubuntu 14.04.2 LTS"
$ gluster --version
glusterfs 3.6.3 built on Jun 12 2015 17:41:05
Repository revision: git://git.gluster.com/glusterfs.git
Copyright (c) 2006-2011 Gluster Inc. http://www.gluster.com
GlusterFS comes with ABSOLUTELY NO WARRANTY.
You may redistribute copies of GlusterFS under the terms of the GNU General Public License.
Fedora:
$ cat /etc/fedora-release
Fedora release 22 (Twenty Two)
$ gluster --version
glusterfs 3.6.3 built on May 5 2015 14:18:23
Repository revision: git://git.gluster.com/glusterfs.git
Copyright (c) 2006-2011 Gluster Inc. http://www.gluster.com
GlusterFS comes with ABSOLUTELY NO WARRANTY.
You may redistribute copies of GlusterFS under the terms of the GNU General Public License.
When running gstatus -a -o json
I am getting non-json data.
Here's the example line causing the aforementioned issue:
gstatus/gstatus/glusterlib/cluster.py
Line 185 in 6cc8935
It should print notes, errors, etc. to stderr instead of stdout, example:
print("error1", file=sys.stderr)
Hi everyone,
Just as info.. also with latest gstatus and gluster 10.1 once inside an replication 1 brick is down > then other Gluster nodes getting an error. even if all is fine (apart of the 1 Brick down). As long the Volume is up > means is not emergency.
may an easy fix?
Traceback (most recent call last):
File "/usr/lib64/python3.6/runpy.py", line 193, in _run_module_as_main
"main", mod_spec)
File "/usr/lib64/python3.6/runpy.py", line 85, in _run_code
exec(code, run_globals)
File "/usr/bin/gstatus/main.py", line 77, in
File "/usr/bin/gstatus/main.py", line 73, in main
File "/usr/bin/gstatus/glusterlib/cluster.py", line 37, in gather_data
File "/usr/bin/gstatus/glustercli/cli/volume.py", line 189, in status_detail
File "/usr/bin/gstatus/glustercli/cli/parsers.py", line 307, in parse_volume_status
AttributeError: 'NoneType' object has no attribute 'get'
thx
Hi ,
this tools is amazing.
easy to use and it gives the most important infos - without to have massive overhead. simple.
what i recognized is,.... or i may doing something wrong.
/usr/bin/gstatus -ab -o keyvalue
brick_count=3,bricks_active=3,glfs_version='3.12.15',node_count=3,nodes_active=3,over_commit='No',product_name='Community',raw_capacity=5736648228864,sh_active=3,sh_enabled=3,snapshot_count=0,status='healthy',usable_capacity=1912216076288,used_capacity=1266406785024,volume_count=1,volume_summary_down=0,volume_summary_degraded=0,volume_summary_partial=0,volume_summary_up=1
it doesn't show up if Files are all in Sync or not.
Seems if any split-brain or file unhealed,.... it's not recognize it.
UP - 3/3 bricks up - Replicate
Capacity: (22% used) 393.00 GiB/1.70 TiB (used/total)
Snapshots: 0
Self Heal: 3/ 3 Heal backlog of 28 files
Tasks Active: None
Protocols: glusterfs:on NFS:on SMB:on
Gluster Connectivty: 3 hosts, 24 tcp connections
If Gluster is in Sync/Heal Mode > then it's not really a Healthy state.
Normally it shows up
Self Heal: 3/ 3 All files in sync
Would this somehow be able to implement,.. so that this is easy to pull from keyvalue? i use icinga to get status of it. not sure if it's recognize Split-brain as well?
thx
Max
Hi, I would like to use gstatus to generate a json that is delivered via a local webserver. When I start gstatus manually with the -o option, the output as json can be generated.
If I start gstatus in a one line script via crontab, the json can not be generated (file size = 0B).
My proxmox host then sends me the following error by email:
/usr/local/bin/gstatus -a -o json > /tmp/test.json
Any Ideas?
Thanks and regards,
Thorsten
in gstatus/libgluster/volume.py
:
self.raw_capacity
is calculated for all bricks, data or arbiter. The gluster vol info --xml
output has an <isArbiter>
flag for each brick, which can easily be used to identify data bricks vs. arbiter bricks.
self.raw_used
is calculated similarly, so the same adjustment should be made.
self.replicaCount
is taken directly from the gluster vol info --xml
output, but for arbiter volumes that count includes the arbiter brick. The XML also provides an <arbiterCount>
value, which should be subtracted from the <replicaCount>
value to give the correct data replica count.
With the above changes made, the self.usable_capacity
and self.used_capacity
calculations should then provide correct values.
The gstatus command failing if gluster command takes more time to get xml data:
...
...
Gluster_Command. Response from glusterd has exceeded the 240 secs timeout, terminating the request
Traceback (most recent call last):
File "/usr/bin/gstatus", line 221, in
main()
File "/usr/bin/gstatus", line 135, in main
cluster.update_state(self_heal_backlog)
File "/usr/lib/python2.7/site-packages/gstatus/libgluster/cluster.py", line 638, in update_state
self.calc_connections()
File "/usr/lib/python2.7/site-packages/gstatus/libgluster/cluster.py", line 730, in calc_connections
xml_root = ETree.fromstring(xml_string)
File "/usr/lib64/python2.7/xml/etree/ElementTree.py", line 1301, in XML
return parser.close()
File "/usr/lib64/python2.7/xml/etree/ElementTree.py", line 1654, in close
self._raiseerror(v)
File "/usr/lib64/python2.7/xml/etree/ElementTree.py", line 1506, in _raiseerror
raise err
xml.etree.ElementTree.ParseError: no element found: line 1, column 0
It seems gstatus does not comply with specified timeout and returns before timeout (within 5 seconds).
Sometimes gluster command takes more time to finish which resulting gstatus failing with above ParseError exception.
Hi - like the utility and it's being used in a Zabbix monitoring project. Wondering if it's possible to also pull in the Gluster quota limits on a volume? Thanks.
Hi,
please fix this Version compatibility problem.
#gstatus -a -o json gstatus is not compatible with this version of glusterfs 4.1.2
I'm using the latest git version.
Thank you!
I'm running gstatus in the latest gluster centos container image.
It seems that the gluster volume heal <VOL> info
command is only available for replicate/disperse volumes:
# gluster vol heal gv0 info
Volume gv0 is not of type replicate/disperse
Volume heal failed.
Is it possible to allow gstatus to skip displaying self-heal info if not present (for example I added a replicate volume (gv1), but gstatus will throw an error because of gv0 is a distribute volume type? This currently blocks us from using gstatus at all since the -a
, -b
and -v
flags are all throwing the same error from /gstatus/glusterlib/display_status.py
.
When attempting to view the gstatus -a
of my cluster, I get the following traceback:
Note: Unable to get self-heal status for one or more volumes
Traceback (most recent call last):
File "/usr/lib64/python3.6/runpy.py", line 193, in _run_module_as_main
"__main__", mod_spec)
File "/usr/lib64/python3.6/runpy.py", line 85, in _run_code
exec(code, run_globals)
File "/usr/local/bin/gstatus/__main__.py", line 74, in <module>
File "/usr/local/bin/gstatus/__main__.py", line 71, in main
File "/usr/local/bin/gstatus/glusterlib/display_status.py", line 11, in display_status
File "/usr/local/bin/gstatus/glusterlib/display_status.py", line 58, in _build_status
KeyError: 'healinfo'
My volume info:
Volume Name: gv0
Type: Distribute
Volume ID: 1360cf07-5a64-4452-aedc-9d0d8aba1280
Status: Started
Snapshot Count: 0
Number of Bricks: 1
Transport-type: tcp
Bricks:
Brick1: gluster-node-1:/export
Options Reconfigured:
nfs.disable: on
storage.fips-mode-rchecksum: on
transport.address-family: inet
Gstatus version:
# gstatus --version
gstatus 1.0.4
Gluster version:
# gluster --version
glusterfs 7.9
Repository revision: git://git.gluster.org/glusterfs.git
Copyright (c) 2006-2016 Red Hat, Inc. <https://www.gluster.org/>
GlusterFS comes with ABSOLUTELY NO WARRANTY.
It is licensed to you under your choice of the GNU Lesser
General Public License, version 3 or any later version (LGPLv3
or later), or the GNU General Public License, version 2 (GPLv2),
in all cases as published by the Free Software Foundation.
I have some rpms build from 8.0 source and it seems that gstatus fails to work:
RPMS:
[root@glustera gstatus]# rpm -qa | grep -E "gluster|python" | sort
glusterfs-8.0-0.0.el8.x86_64
glusterfs-cli-8.0-0.0.el8.x86_64
glusterfs-client-xlators-8.0-0.0.el8.x86_64
glusterfs-fuse-8.0-0.0.el8.x86_64
glusterfs-gnfs-8.0-0.0.el8.x86_64
glusterfs-server-8.0-0.0.el8.x86_64
libglusterd0-8.0-0.0.el8.x86_64
libglusterfs0-8.0-0.0.el8.x86_64
platform-python-3.6.8-23.el8.x86_64
platform-python-pip-9.0.3-16.el8.noarch
platform-python-setuptools-39.2.0-5.el8.noarch
policycoreutils-python-utils-2.9-9.el8.noarch
python2-2.7.17-1.module_el8.2.0+381+9a5b3c3b.x86_64
python2-libs-2.7.17-1.module_el8.2.0+381+9a5b3c3b.x86_64
python2-pip-9.0.3-16.module_el8.2.0+381+9a5b3c3b.noarch
python2-pip-wheel-9.0.3-16.module_el8.2.0+381+9a5b3c3b.noarch
python2-setuptools-39.0.1-11.module_el8.2.0+381+9a5b3c3b.noarch
python2-setuptools-wheel-39.0.1-11.module_el8.2.0+381+9a5b3c3b.noarch
python36-3.6.8-2.module_el8.1.0+245+c39af44f.x86_64
python3-audit-3.0-0.17.20191104git1c2f876.el8.x86_64
python3-cairo-1.16.3-6.el8.x86_64
python3-configobj-5.0.6-11.el8.noarch
python3-dateutil-2.6.1-6.el8.noarch
python3-dbus-1.2.4-15.el8.x86_64
python3-decorator-4.2.1-2.el8.noarch
python3-dmidecode-3.12.2-15.el8.x86_64
python3-dnf-4.2.17-6.el8.noarch
python3-dnf-plugins-core-4.0.12-3.el8.noarch
python3-firewall-0.8.0-4.el8.noarch
python3-gobject-3.28.3-1.el8.x86_64
python3-gobject-base-3.28.3-1.el8.x86_64
python3-gpg-1.10.0-6.el8.0.1.x86_64
python3-hawkey-0.39.1-5.el8.x86_64
python3-libcomps-0.1.11-4.el8.x86_64
python3-libdnf-0.39.1-5.el8.x86_64
python3-libs-3.6.8-23.el8.x86_64
python3-libselinux-2.9-3.el8.x86_64
python3-libsemanage-2.9-2.el8.x86_64
python3-libxml2-2.9.7-7.el8.x86_64
python3-linux-procfs-0.6-7.el8.noarch
python3-nftables-0.9.3-12.el8.x86_64
python3-perf-4.18.0-193.14.2.el8_2.x86_64
python3-pip-9.0.3-16.el8.noarch
python3-pip-wheel-9.0.3-16.el8.noarch
python3-policycoreutils-2.9-9.el8.noarch
python3-pyudev-0.21.0-7.el8.noarch
python3-pyxattr-0.5.3-18.el8.x86_64
python3-rpm-4.14.2-37.el8.x86_64
python3-schedutils-0.6-6.el8.x86_64
python3-setools-4.2.2-2.el8.x86_64
python3-setuptools-39.2.0-5.el8.noarch
python3-setuptools-wheel-39.2.0-5.el8.noarch
python3-six-1.11.0-8.el8.noarch
python3-slip-0.6.4-11.el8.noarch
python3-slip-dbus-0.6.4-11.el8.noarch
python3-syspurpose-1.26.17-1.el8_2.x86_64
python3-systemd-234-8.el8.x86_64
samba-vfs-glusterfs-4.11.11-1.el8.x86_64
Python2 fails with:
[root@glustera gstatus]# /usr/bin/python2.7 ./gstatus.py
Traceback (most recent call last):
File "./gstatus.py", line 244, in <module>
main()
File "./gstatus.py", line 132, in main
cluster.initialise()
File "/root/gstatus/gstatus/libgluster/cluster.py", line 97, in initialise
self.define_nodes()
File "/root/gstatus/gstatus/libgluster/cluster.py", line 170, in define_nodes
local_ip_list = get_ipv4_addr() # Grab all IP's
File "/root/gstatus/gstatus/libutils/network.py", line 130, in get_ipv4_addr
namestr = names.tobytes()
AttributeError: 'array.array' object has no attribute 'tobytes'
Python3 fails with:
Traceback (most recent call last):bmeta
File "./gstatus.py", line 142, in main
cluster.update_state(self_heal_backlog, client_status)
File "/root/gstatus/gstatus/libgluster/cluster.py", line 564, in update_state
task.status_str = task.find('./statusStr').text
AttributeError: 'xml.etree.ElementTree.Element' object has no attribute 'status_str'
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "./gstatus.py", line 244, in <module>
main()
File "./gstatus.py", line 143, in main
except [GlusterFailedVolume, GlusterNotPeerNode] as e:
TypeError: catching classes that do not inherit from BaseException is not allowed
Gluster is operational:
[root@glustera gstatus]# gluster pool list
UUID Hostname State
9ff2f29c-78a0-4490-9c60-32a752ffd80a glusterb Connected
45854556-04d2-4c62-937f-326ca2cb8677 glusterc Connected
25e3c9b9-17e0-43c2-bd31-276381a4450f glusterd Connected
dd6d1704-7049-4748-9916-b66b748656bb glustere Connected
1f699dfd-15f8-4f14-83c0-62b695fad7e0 localhost Connected
[root@glustera gstatus]# gluster volume list
ctdbmeta
custdata
The problem is in this line
https://github.com/gluster/gstatus/blob/master/gstatus/libutils/utils.py#L65
The test fails with minor version (0<4) although major version passes (4>3).
It should be rewritted to
if (int(this_major) > int(tgt_major)):
return True
elif (int(this_major) == int(tgt_major)):
if (int(this_minor) >= int(tgt_minor)):
return True
else:
return False
else:
return False
Hi,
I'm trying to install gstatus from source via pip install .
I get the following error:
Traceback (most recent call last):
File "<string>", line 1, in <module>
File "/tmp/pip-req-build-iu5z4zm1/setup.py", line 7, in <module>
from gstatus import version
ImportError: cannot import name 'version' from 'gstatus' (/tmp/pip-req-build-iu5z4zm1/gstatus/__init__.py)
Can this version file be included in the package?
Hi, no matter what i do, either using source or binary version, always getting this error:
sudo gstatus -v
Traceback (most recent call last):
File "/usr/local/lib/python3.8/dist-packages/glustercli/cli/parsers.py", line 218, in parse_volume_info
volumes.append(_parse_a_vol(volume_el))
File "/usr/local/lib/python3.8/dist-packages/glustercli/cli/parsers.py", line 137, in _parse_a_vol
'stripe': int(volume_el.find('stripeCount').text),
AttributeError: 'NoneType' object has no attribute 'text'
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/usr/local/bin/gstatus", line 11, in
load_entry_point('gstatus==1.0.6', 'console_scripts', 'gstatus')()
File "/usr/local/lib/python3.8/dist-packages/gstatus-1.0.6-py3.8.egg/gstatus/main.py", line 73, in main
File "/usr/local/lib/python3.8/dist-packages/gstatus-1.0.6-py3.8.egg/gstatus/glusterlib/cluster.py", line 37, in gather_data
File "/usr/local/lib/python3.8/dist-packages/glustercli/cli/volume.py", line 188, in status_detail
info(volname),
File "/usr/local/lib/python3.8/dist-packages/glustercli/cli/volume.py", line 152, in info
return parse_volume_info(volume_execute_xml(cmd),
File "/usr/local/lib/python3.8/dist-packages/glustercli/cli/parsers.py", line 220, in parse_volume_info
raise GlusterCmdOutputParseError(err)
glustercli.cli.parsers.GlusterCmdOutputParseError: 'NoneType' object has no attribute 'text'
Running gstatus script throws a DeprecationWarning
/usr/bin/gstatus/main.py:7: DeprecationWarning: The distutils package is deprecated and slated for removal in Python 3.12. Use setuptools or check PEP 632 for potential alternatives
gstatus --version : gstatus v1.0.8
Dist : Ubuntu 22.04.1 LTS
python3 version : Python 3.10.4
when running gstatus -a -o json
to get gluster json
it return my object field number values as strings (type)
example object:
"healinfo": [
{
"name": "device:/rhgs/node1_brick1/disk_vol2",
"status": "Connected",
"host_uuid": "7c99b0ca-d912-4fc7-b4c0-535686744dc3",
"nr_entries": "21" >> this value "21" is returned with type of "String" instead of "Integer"
we would like to get it as Intger value
however some of them returns the "-" sign instead of actual values
I'm running the gstatus package for CentOS 8 from COPR with glusterfs 7.6, and with bricks missing it hits an exception:
# gstatus -a
Traceback (most recent call last):
File "/usr/bin/gstatus", line 245, in <module>
main()
File "/usr/bin/gstatus", line 143, in main
cluster.update_state(self_heal_backlog, client_status)
File "/usr/lib/python2.7/site-packages/gstatus/libgluster/cluster.py", line 531, in update_state
self.volume[volume_name].update(xml_obj)
File "/usr/lib/python2.7/site-packages/gstatus/libgluster/volume.py", line 128, in update
self.brick_update(node)
File "/usr/lib/python2.7/site-packages/gstatus/libgluster/volume.py", line 99, in brick_update
node_info['fsName'],
KeyError: 'fsName'
The drive was in an i/o error status when this exception was occurring, if that's relevant:
# ll /data/glusterfs/volname1/
ls: cannot access '/data/glusterfs/volname1/brick2': Input/output error
total 0
drwxr-xr-x. 5 root root 70 Feb 17 11:59 brick1
d?????????? ? ? ? ? ? brick2
Hi!
The latest gstatus commit removes an if: before main() but not the else: which leads to a syntax error.
Hey, in former versions there was an outputmode keyvalue which was easiely to use with nagios and such ... Could that be re-added please?
I'm getting error when trying to run gstatus on nodes[2:3] of the cluster. It only works on node1.
[root@gluster2 ~]# gstatus -a -o json
Unable to associate brick gluster1.mydomain.com:/bricks/brick1/data with a peer in the cluster, possibly due
to name lookup failures. If the nodes are not registered (fwd & rev)
to dns, add local entries for your cluster nodes in the the /etc/hosts file
Here is my peer status:
[root@gluster2 ~]# gluster peer status
Number of Peers: 2
Hostname: 10.10.10.5
Uuid: bbb19d4b-c9fc-4f14-8a9c-9cedf1fac367
State: Peer in Cluster (Connected)
Other names:
gluster1.mydomain.com
Hostname: gluster3.mydomain.com
Uuid: da74fc61-eb2a-4f0d-906b-9fc2c287908f
State: Peer in Cluster (Connected)
As you may notice my node1 hostname is IP address. According to documentation:
Note: When using hostnames, the first server needs to be probed from one other server to set its hostname.
I did that, and the hostname appeared in Other names
And here is my gluster volume config:
[root@gluster2 ~]# gluster volume info
Volume Name: data
Type: Replicate
Volume ID: 8f8285bc-c8b2-4112-b88e-49df45bf7e87
Status: Started
Snapshot Count: 0
Number of Bricks: 1 x 3 = 3
Transport-type: tcp
Bricks:
Brick1: gluster1.mydomain.com:/bricks/brick1/data
Brick2: gluster2.mydomain.com:/bricks/brick1/data
Brick3: gluster3.mydomain.com:/bricks/brick1/data
Options Reconfigured:
transport.address-family: inet
nfs.disable: on
performance.client-io-threads: off
Hello, im using gstatus v1 and there is no option for output... that is a bug or just -o became depracated?
[root@aaaa~]# gstatus --version
gstatus 1.0.0
[root@aaaa~]gstatus -o json
Usage: gstatus [options]gstatus: error: no such option: -o
[root@aaaa~]# gstatus -h
Usage: gstatus [options]Options:
--version show program's version number and exit
-h, --help show this help message and exit
-v, --volume Supply a volume or a list of volumes by repeated
invocation of -v. A regular expression can be provided
in place of a volume name (ensure to use single quotes
around the expression)
-a, --all Print all available details on volumes
-b, --bricks Print the list of bricks
-q, --quota Print the quota information
-s, --snapshots Print the snapshot information
-u UNITS, --units=UNITS
display storage size in given units
Hello!
Capacity values seem nonsensical to me.. See below:
`# gstatus --all
Cluster:
Status: Healthy GlusterFS: 10.3
Nodes: 5/5 Volumes: 5/5
Volumes:
backups
Replicate Started (UP) - 5/5 Bricks Up
Capacity: (86.62% used) 3.00 TiB/4.00 TiB (used/total)
Bricks:
Distribute Group 1:
10.100.13.71:/data/brick2/backups (Online)
10.100.13.73:/data/brick2/backups (Online)
10.100.13.72:/data/brick2/backups (Online)
10.100.13.75:/data/brick1/backups (Online)
10.100.13.74:/data/brick2/backups (Online)
engine
Replicate Started (UP) - 5/5 Bricks Up
Capacity: (557.29% used) 3.00 TiB/634.00 GiB (used/total)
Bricks:
Distribute Group 1:
10.100.13.72:/data/brick1/engine (Online)
10.100.13.71:/data/brick1/engine (Online)
10.100.13.73:/data/brick1/engine (Online)
10.100.13.75:/data/brick1/engine (Online)
10.100.13.74:/data/brick1/engine (Online)
guests
Replicate Started (UP) - 5/5 Bricks Up
Capacity: (557.29% used) 3.00 TiB/634.00 GiB (used/total)
Bricks:
Distribute Group 1:
10.100.13.71:/data/brick1/guests (Online)
10.100.13.73:/data/brick1/guests (Online)
10.100.13.72:/data/brick1/guests (Online)
10.100.13.75:/data/brick1/guests (Online)
10.100.13.74:/data/brick1/guests (Online)
kotidata
Replicate Started (UP) - 5/5 Bricks Up
Capacity: (86.62% used) 3.00 TiB/4.00 TiB (used/total)
Bricks:
Distribute Group 1:
10.100.13.71:/data/brick2/kotidata (Online)
10.100.13.73:/data/brick2/kotidata (Online)
10.100.13.72:/data/brick2/kotidata (Online)
10.100.13.75:/data/brick1/kotidata (Online)
10.100.13.74:/data/brick2/kotidata (Online)
nvr
Replicate Started (UP) - 2/2 Bricks Up
Capacity: (544.19% used) 3.00 TiB/650.00 GiB (used/total)
Bricks:
Distribute Group 1:
10.100.13.75:/data/brick1/nvr (Online)
10.100.13.74:/data/brick1/nvr (Online)
`
Currently works only on python2, prioritize to port to python3.
Please, make a new release of the code, with new version, new tag and new tarball for that new tag.
Hello,
Today I upgraded glusterfs to 7.0 ( yay ).
However, after the upgrade I noticed that gstatus is not working any more :(
Is there any plan to make a release to gstatus in order to support gluster 7?
Brick demo1:/gluster/bricks/1/brick 49152 0 Y 1846
Brick demo2:/gluster/bricks/2/brick 49153 0 Y 24646
Brick arbiter:/gluster/bricks/3/brick 49152 0 Y 16762
Self-heal Daemon on localhost N/A N/A Y 26784
Self-heal Daemon on demo1 N/A N/A Y 1931
Self-heal Daemon on arbiter N/A N/A Y 7603
There are no active volume tasks
root@demo2:~# gstatus -a
gstatus is not compatible with this version of glusterfs 7.0
Thanks!
Hi, I have a single node gluster server, as setup by oVirt in single node hyperconverged mode.
When running gstatus 0.66, I get :
~]# gstatus --all
Unable to associate brick server.fqdn:/gluster_bricks/engine/engine with a peer in the cluster, possibly dueto name lookup failures. If the nodes are not registered (fwd & rev)to dns, add local entries for your cluster nodes in the the /etc/hosts file
No such error on a 3 node cluster
A declarative, efficient, and flexible JavaScript library for building user interfaces.
๐ Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
An Open Source Machine Learning Framework for Everyone
The Web framework for perfectionists with deadlines.
A PHP framework for web artisans
Bring data to life with SVG, Canvas and HTML. ๐๐๐
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
Some thing interesting about web. New door for the world.
A server is a program made to process requests and deliver data to clients.
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
Some thing interesting about visualization, use data art
Some thing interesting about game, make everyone happy.
We are working to build community through open source technology. NB: members must have two-factor auth.
Open source projects and samples from Microsoft.
Google โค๏ธ Open Source for everyone.
Alibaba Open Source for everyone
Data-Driven Documents codes.
China tencent open source team.