2
0
mirror of https://github.com/munin-monitoring/contrib.git synced 2018-11-08 00:59:34 +01:00
contrib-munin/plugins/other/moinoin_pages

114 lines
3.9 KiB
Plaintext
Raw Normal View History

2007-08-24 02:28:13 +02:00
#! /usr/local/bin/python
# Overview
# --------
#
# this is a munin plugin that lists the number of pages (including ACL-protected pages) in all wikis of a MoinMoin wikifarm
#
# Installation
# ------------
#
# Put this in your munin plugins directory. You probably want to set the following block in plugin-conf.d/munin-node so that it runs under the riht user for your wiki:
#
# [moinmoin_*]
# user www
#
# Implementation notes
# --------------------
#
# it is quite koumbit-specific:
# 1. the wikifarm config is hardcoded
# 2. it relies on the "wikilist.py" file to contain the list of wiki -> url patterns
# 3. it assumes the url patterns are simple enough that they are decodable into an url
#
# also note that it reuses code from MoinMoin/wikimacro.py's SystemInfo macro
#
# finally, i tried using XMLRPC instead of native functions to fetch the data, but it ended up being slower. For the record, here is what the getPageList() call would have looked like:
# xmlrpclib.ServerProxy("http://wiki.koumbit.net/?action=xmlrpc2").getAllPages()
#
# the quick benchmark i did yieled those results for the getAllPages() vs getPageList() calls:
# xmlrpc: 2.35 real 0.12 user 0.04 sys
# native: 1.44 real 1.07 user 0.35 sys
#
# so the plugin is spending more time in the CPU (all time, actually), but it's doing in faster. it is highly possible that the CPU time spared in XMLRPC is in fact used by the server
#
# (C) Copyleft 2007, The Anarcat <anarcat@koumbit.org>
# Licensed under the GPLv2 or any later version
import sys, operator, os
os.chdir('/export/wiki/config')
sys.path.insert(0, '/export/wiki/config')
from MoinMoin import wikiutil
from MoinMoin.Page import Page
from farmconfig import wikis
from re import sub
import farmconfig
from MoinMoin.request import RequestCLI
def _formatInReadableUnits(size):
size = float(size)
unit = u' Byte'
if size > 9999:
unit = u' KiB'
size /= 1024
if size > 9999:
unit = u' MiB'
size /= 1024
if size > 9999:
unit = u' GiB'
size /= 1024
return u"%.1f %s" % (size, unit)
def _getDirectorySize(path):
try:
dirsize = 0
for root, dirs, files in os.walk(path):
dirsize += sum([os.path.getsize(os.path.join(root, name)) for name in files])
except EnvironmentError, e:
dirsize = -1
return dirsize
def main():
for wiki in wikis:
name = wiki[0]
url = wiki[1]
# XXX, hack: transform the regexp into a canonical url
# we need canonical urls in the config for this to be clean
# look for (foo|bar) and replace with foo
url = sub('\(([^\|]*)(\|[^\)]*\))+', '\\1', url)
# remove common regexp patterns and slap a protocol to make this a real url
url = sub('[\^\$]|(\.\*)', '', url)
mod = getattr(__import__(name), 'Config')
#print "Upgradeing wiki %s (%s)" % (getattr(mod, 'sitename'), url)
request = RequestCLI(url)
pagelist = request.rootpage.getPageList(user='')
systemPages = [page for page in pagelist
if wikiutil.isSystemPage(request, page)]
print(name + '.value ' + str(len(pagelist)-len(systemPages)))
#totalsize = reduce(operator.add, [Page(request, name).size() for name in pagelist])
#print('Accumulated page sizes' + _formatInReadableUnits(totalsize))
def config():
print("""graph_title Wiki size
graph_vlabel Number of pages
graph_args --base 1000 -l 0
graph_scale no
graph_category Wiki
graph_info The number of pages excludes system pages but includes ACL-protected pages.""")
for wiki in wikis:
name = wiki[0]
mod = getattr(__import__(name), 'Config')
print(name + '.label ' + getattr(mod, 'sitename'))
if __name__ == "__main__":
if len(sys.argv) > 1 and sys.argv[1] == 'config':
config()
else:
main()