/usr/share/webcheck/plugins/size.py is in webcheck 1.10.4.
This file is owned by root:root, with mode 0o644.
The actual contents of the file can be viewed below.
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 | # size.py - plugin that lists pages that could be slow to load
#
# Copyright (C) 1998, 1999 Albert Hopkins (marduk)
# Copyright (C) 2002 Mike W. Meyer
# Copyright (C) 2005, 2006 Arthur de Jong
#
# This program is free software; you can redistribute it and/or modify
# it under the terms of the GNU General Public License as published by
# the Free Software Foundation; either version 2 of the License, or
# (at your option) any later version.
#
# This program is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU General Public License for more details.
#
# You should have received a copy of the GNU General Public License
# along with this program; if not, write to the Free Software
# Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA
#
# The files produced as output from the software do not automatically fall
# under the copyright of the software, unless explicitly stated otherwise.
"""Present a list of pages that are large and probably slow to download."""
__title__ = "what's big"
__author__ = 'Arthur de Jong'
__outputfile__ = 'size.html'
import config
import plugins
def _getsize(link, done=None):
"""Return the size of the link and all its embedded links, counting each
link only once."""
# make a new list
if done is None:
done = []
# add this link to the list
done.append(link)
# if we don't known about our total size yet, calculate
if not hasattr(link, 'totalSize'):
size = 0
# add our size
if link.size is not None:
size = link.size
# add sizes of embedded objects
for embed in link.embedded:
if embed not in done:
size += _getsize(embed, done)
link.totalSize = size
return link.totalSize
def generate(site):
"""Output the list of large pages to the given file descriptor."""
# get all internal pages and get big links
links = [ x
for x in site.linkMap.values()
if x.ispage and
x.isinternal and
_getsize(x) >= config.REPORT_SLOW_URL_SIZE*1024 ]
# sort links by size (biggest first)
links.sort(lambda a, b: cmp(b.totalSize, a.totalSize))
# present results
fp = plugins.open_html(plugins.size, site)
if not links:
fp.write(
' <p class="description">\n'
' No pages over %(size)dK were found.\n'
' </p>\n'
% { 'size': config.REPORT_SLOW_URL_SIZE })
plugins.close_html(fp)
return
fp.write(
' <p class="description">\n'
' These pages are probably too big (over %(size)dK) which could be\n'
' slow to download.\n'
' </p>\n'
' <ul>\n'
% { 'size': config.REPORT_SLOW_URL_SIZE })
for link in links:
size = plugins.get_size(link.totalSize)
fp.write(
' <li>\n'
' %(link)s\n'
' <ul class="problem">\n'
' <li>size: %(size)s</li>\n'
' </ul>\n'
' </li>\n'
% { 'link': plugins.make_link(link),
'size': size })
link.add_pageproblem(
'this page and its components is %(size)s'
% { 'size': size })
fp.write(
' </ul>\n' )
plugins.close_html(fp)
|