Files
server/usr/lib/python3/dist-packages/mod_python/apache.py
2026-01-07 20:52:11 +01:00

1205 lines
38 KiB
Python

# vim: set sw=4 expandtab :
#
# Copyright (C) 2000, 2001, 2013 Gregory Trubetskoy
# Copyright (C) 2002, 2003, 2004, 2005, 2006, 2007 Apache Software Foundation
#
# Licensed under the Apache License, Version 2.0 (the "License"); you
# may not use this file except in compliance with the License. You
# may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or
# implied. See the License for the specific language governing
# permissions and limitations under the License.
#
# Originally developed by Gregory Trubetskoy.
#
import sys
import traceback
import time
import os
import pdb
import stat
import inspect
import importlib
import types
import _apache
try:
from html import escape
except:
from cgi import escape
try:
import threading
except:
import dummy_threading as threading
# Cache for values of PythonPath that have been seen already.
_path_cache = {}
_path_cache_lock = threading.Lock()
_result_warning = """Handler has returned result or raised SERVER_RETURN
exception with argument having non integer type. Type of value returned
was %s, whereas expected """ + str(int) + "."
class CallBack:
"""
A generic callback object.
"""
class HStack:
"""
The actual stack string lives in the request object so
it can be manipulated by both apache.py and mod_python.c
"""
def __init__(self, req):
self.req = req
def pop(self):
handlers = self.req.hstack.split()
if not handlers:
return None
else:
self.req.hstack = " ".join(handlers[1:])
return handlers[0]
def ConnectionDispatch(self, conn):
# config
config, debug = conn.base_server.get_config(), False
if "PythonDebug" in config:
debug = config["PythonDebug"] == "1"
try:
handler = conn.hlist.handler
# split module::handler
l = handler.split('::', 1)
module_name = l[0]
if len(l) == 1:
# no oject, provide default
obj_str = "connectionhandler"
else:
obj_str = l[1]
# evaluate pythonpath and set sys.path to
# resulting value if not already done
if "PythonPath" in config:
_path_cache_lock.acquire()
try:
pathstring = config["PythonPath"]
if pathstring not in _path_cache:
newpath = eval(pathstring)
_path_cache[pathstring] = None
sys.path[:] = newpath
finally:
_path_cache_lock.release()
# import module
autoreload = True
if "PythonAutoReload" in config:
autoreload = config["PythonAutoReload"] == "1"
module = import_module(module_name,
autoreload=autoreload,
log=debug)
# find the object
obj = resolve_object(module, obj_str,
arg=conn, silent=0)
# Only permit debugging using pdb if Apache has
# actually been started in single process mode.
pdb_debug = False
if "PythonEnablePdb" in config:
pdb_debug = config["PythonEnablePdb"] == "1"
if pdb_debug and exists_config_define("ONE_PROCESS"):
# Don't use pdb.runcall() as it results in
# a bogus 'None' response when pdb session
# is quit. With this code the exception
# marking that the session has been quit is
# propogated back up and it is obvious in
# the error message what actually occurred.
debugger = pdb.Pdb()
debugger.reset()
sys.settrace(debugger.trace_dispatch)
try:
result = obj(conn)
finally:
debugger.quitting = 1
sys.settrace(None)
else:
result = obj(conn)
assert (result.__class__ is int), \
"ConnectionHandler '%s' returned invalid return code." % handler
except:
# Error (usually parsing)
try:
exc_type, exc_value, exc_traceback = sys.exc_info()
result = self.ReportError(exc_type, exc_value, exc_traceback, srv=conn.base_server,
phase="ConnectionHandler",
hname=handler, debug=debug)
finally:
exc_traceback = None
return result
def FilterDispatch(self, fltr):
req = fltr.req
# config
config, debug = req.get_config(), False
if "PythonDebug" in config:
debug = config["PythonDebug"] == "1"
try:
# split module::handler
l = fltr.handler.split('::', 1)
module_name = l[0]
if len(l) == 1:
# no oject, provide default
if fltr.is_input:
obj_str = "inputfilter"
else:
obj_str = "outputfilter"
else:
obj_str = l[1]
# add the directory to pythonpath if
# not there yet, or evaluate pythonpath
# and set sys.path to resulting value
# if not already done
if "PythonPath" in config:
_path_cache_lock.acquire()
try:
pathstring = config["PythonPath"]
if pathstring not in _path_cache:
newpath = eval(pathstring)
_path_cache[pathstring] = None
sys.path[:] = newpath
finally:
_path_cache_lock.release()
else:
if fltr.dir:
_path_cache_lock.acquire()
try:
if fltr.dir not in sys.path:
sys.path[:0] = [fltr.dir]
finally:
_path_cache_lock.release()
# import module
autoreload = True
if "PythonAutoReload" in config:
autoreload = config["PythonAutoReload"] == "1"
module = import_module(module_name,
autoreload=autoreload,
log=debug)
# find the object
obj = resolve_object(module, obj_str,
arg=fltr, silent=0)
# Only permit debugging using pdb if Apache has
# actually been started in single process mode.
pdb_debug = False
if "PythonEnablePdb" in config:
pdb_debug = config["PythonEnablePdb"] == "1"
if pdb_debug and exists_config_define("ONE_PROCESS"):
# Don't use pdb.runcall() as it results in
# a bogus 'None' response when pdb session
# is quit. With this code the exception
# marking that the session has been quit is
# propogated back up and it is obvious in
# the error message what actually occurred.
debugger = pdb.Pdb()
debugger.reset()
sys.settrace(debugger.trace_dispatch)
try:
result = obj(fltr)
finally:
debugger.quitting = 1
sys.settrace(None)
else:
result = obj(fltr)
# always flush the filter. without a FLUSH or EOS bucket,
# the content is never written to the network.
# XXX an alternative is to tell the user to flush() always
if not fltr.closed:
fltr.flush()
except SERVER_RETURN as value:
# SERVER_RETURN indicates a non-local abort from below
# with value as (result, status) or (result, None) or result
try:
if len(value.args) == 2:
(result, status) = value.args
if status:
req.status = status
else:
result = value.args[0]
if result.__class__ is not int:
s = "Value raised with SERVER_RETURN is invalid. It is a "
s = s + "%s, but it must be a tuple or an int." % result.__class__
_apache.log_error(s, APLOG_ERR, req.server)
return
except:
pass
except:
# Error (usually parsing)
try:
exc_type, exc_value, exc_traceback = sys.exc_info()
fltr.disable()
result = self.ReportError(exc_type, exc_value, exc_traceback, req=req, filter=fltr,
phase=fltr.name, hname=fltr.handler,
debug=debug)
finally:
exc_traceback = None
return OK
def HandlerDispatch(self, req):
"""
This is the handler dispatcher.
"""
# be cautious
result = HTTP_INTERNAL_SERVER_ERROR
# config
config, debug = req.get_config(), False
if "PythonDebug" in config:
debug = config["PythonDebug"] == "1"
default_obj_str = _phase_handler_names[req.phase]
# Lookup expected status values that allow us to
# continue when multiple handlers exist.
expected = _status_values[default_obj_str]
try:
hlist = req.hlist
while hlist.handler is not None:
# split module::handler
l = hlist.handler.split('::', 1)
module_name = l[0]
if len(l) == 1:
# no object, provide default
obj_str = default_obj_str
else:
obj_str = l[1]
# add the directory to pythonpath if
# not there yet, or evaluate pythonpath
# and set sys.path to resulting value
# if not already done
if "PythonPath" in config:
_path_cache_lock.acquire()
try:
pathstring = config["PythonPath"]
if pathstring not in _path_cache:
newpath = eval(pathstring)
_path_cache[pathstring] = None
sys.path[:] = newpath
finally:
_path_cache_lock.release()
else:
if not hlist.is_location:
directory = hlist.directory
if directory:
_path_cache_lock.acquire()
try:
if directory not in sys.path:
sys.path[:0] = [directory]
finally:
_path_cache_lock.release()
# import module
autoreload = True
if "PythonAutoReload" in config:
autoreload = config["PythonAutoReload"] == "1"
module = import_module(module_name,
autoreload=autoreload,
log=debug)
# find the object
if '.' not in obj_str: # this is an optimization
try:
obj = module.__dict__[obj_str]
except:
if not hlist.silent:
s = "module '%s' contains no '%s'" % (module.__file__, obj_str)
raise AttributeError(s)
else:
obj = resolve_object(module, obj_str,
arg=req, silent=hlist.silent)
if not hlist.silent or obj is not None:
try:
# Only permit debugging using pdb if Apache has
# actually been started in single process mode.
pdb_debug = False
if "PythonEnablePdb" in config:
pdb_debug = config["PythonEnablePdb"] == "1"
if pdb_debug and exists_config_define("ONE_PROCESS"):
# Don't use pdb.runcall() as it results in
# a bogus 'None' response when pdb session
# is quit. With this code the exception
# marking that the session has been quit is
# propogated back up and it is obvious in
# the error message what actually occurred.
debugger = pdb.Pdb()
debugger.reset()
sys.settrace(debugger.trace_dispatch)
try:
result = obj(req)
finally:
debugger.quitting = 1
sys.settrace(None)
else:
result = obj(req)
except SERVER_RETURN as value:
# The SERVER_RETURN exception type when raised
# otherwise indicates an abort from below with
# value as (result, status) or (result, None) or
# result.
if len(value.args) == 2:
(result, status) = value.args
if status:
req.status = status
else:
result = value.args[0]
assert (result.__class__ is int), \
_result_warning % result.__class__
# stop cycling through handlers
if result not in expected:
break
elif hlist.silent:
# A missing handler when in silent mode will
# only propagate DECLINED if it is the first
# and only handler.
if result == HTTP_INTERNAL_SERVER_ERROR:
result = DECLINED
hlist.next()
except:
# Error (usually parsing)
try:
exc_type, exc_value, exc_traceback = sys.exc_info()
result = self.ReportError(exc_type, exc_value, exc_traceback, req=req,
phase=req.phase, hname=hlist.handler, debug=debug)
finally:
exc_traceback = None
return result
def IncludeDispatch(self, fltr, tag, code):
try:
# config
config, debug = fltr.req.get_config(), False
if "PythonDebug" in config:
debug = config["PythonDebug"] == "1"
if not hasattr(fltr.req,"ssi_globals"):
fltr.req.ssi_globals = {}
fltr.req.ssi_globals["filter"] = fltr
fltr.req.ssi_globals["__file__"] = fltr.req.filename
code = code.replace('\r\n', '\n').rstrip()
if tag == 'eval':
result = eval(code, fltr.req.ssi_globals)
if result is not None:
fltr.write(str(result))
elif tag == 'exec':
exec(code, fltr.req.ssi_globals)
fltr.flush()
except:
try:
exc_type, exc_value, exc_traceback = sys.exc_info()
fltr.disable()
result = self.ReportError(exc_type, exc_value, exc_traceback,
req=fltr.req, filter=fltr,
phase=fltr.name,
hname=fltr.req.filename,
debug=debug)
finally:
exc_traceback = None
raise
fltr.req.ssi_globals["filter"] = None
return OK
def ImportDispatch(self, name):
# config
config, debug = main_server.get_config(), False
if "PythonDebug" in config:
debug = config["PythonDebug"] == "1"
# evaluate pythonpath and set sys.path to
# resulting value if not already done
if "PythonPath" in config:
_path_cache_lock.acquire()
try:
pathstring = config["PythonPath"]
if pathstring not in _path_cache:
newpath = eval(pathstring)
_path_cache[pathstring] = None
sys.path[:] = newpath
finally:
_path_cache_lock.release()
# split module::function
l = name.split('::', 1)
module_name = l[0]
func_name = None
if len(l) != 1:
func_name = l[1]
module = import_module(module_name, log=debug)
if func_name:
getattr(module, func_name)()
def ReportError(self, etype, evalue, etb, req=None, filter=None, srv=None,
phase="N/A", hname="N/A", debug=0):
"""
This function is only used when debugging is on.
It sends the output similar to what you'd see
when using Python interactively to the browser
"""
try: # try/finally
try: # try/except
if str(etype) == "exceptions.IOError" \
and str(evalue)[:5] == "Write":
# if this is an IOError while writing to client,
# it is probably better not to try to write to the cleint
# even if debug is on.
debug = 0
# write to log
for e in traceback.format_exception(etype, evalue, etb):
s = "%s %s: %s" % (phase, hname, e[:-1])
if req:
req.log_error(s, APLOG_ERR)
else:
_apache.log_error(s, APLOG_ERR, srv)
if not debug or not req:
return HTTP_INTERNAL_SERVER_ERROR
else:
# write to client
req.status = HTTP_INTERNAL_SERVER_ERROR
req.content_type = 'text/html'
s = '\n<pre>\nMod_python error: "%s %s"\n\n' % (phase, hname)
for e in traceback.format_exception(etype, evalue, etb):
s = s + escape(e) + '\n'
s = s + "</pre>\n"
if filter:
filter.write(s)
filter.flush()
else:
req.write(s)
return DONE
except:
# last try
traceback.print_exc()
sys.stderr.flush()
finally:
# erase the traceback
etb = None
# we do not return anything
def import_module(module_name, autoreload=1, log=0, path=None):
"""
Get the module to handle the request. If
autoreload is on, then the module will be reloaded
if it has changed since the last import.
"""
# nlehuen: this is a big lock, we'll have to refine it later to get better performance.
# For now, we'll concentrate on thread-safety.
# (Re)import
if module_name in sys.modules:
# The module has been imported already
module = sys.modules[module_name]
oldmtime, mtime = 0, 0
if autoreload:
# but is it in the path?
try:
file = module.__dict__["__file__"]
except KeyError:
file = None
# the "and not" part of this condition is to prevent execution
# of arbitrary already imported modules, such as os. The
# reason we use startswith as opposed to exact match is that
# modules inside packages are actually in subdirectories.
if not file or (path and not list(filter(file.startswith, path))):
# there is a script by this name already imported, but it's in
# a different directory, therefore it's a different script
mtime, oldmtime = 0, -1 # trigger import
else:
try:
last_check = module.__dict__["__mtime_check__"]
except KeyError:
last_check = 0
if (time.time() - last_check) > 1:
oldmtime = module.__dict__.get("__mtime__", 0)
mtime = module_mtime(module)
else:
pass
else:
mtime, oldmtime = 0, -1
if mtime != oldmtime:
# Import the module
if log:
if path:
s = "mod_python: (Re)importing module '%s' with path set to '%s'" % (module_name, path)
else:
s = "mod_python: (Re)importing module '%s'" % module_name
_apache.log_error(s, APLOG_NOTICE)
parent = None
module = importlib.import_module(module_name)
if mtime == 0:
mtime = module_mtime(module)
module.__mtime__ = mtime
return module
def module_mtime(module):
"""Get modification time of module"""
mtime = 0
if "__file__" in module.__dict__:
filepath = module.__file__
try:
# this try/except block is a workaround for a Python bug in
# 2.0, 2.1 and 2.1.1. See
# http://sourceforge.net/tracker/?group_id=5470&atid=105470&func=detail&aid=422004
if os.path.exists(filepath):
mtime = os.path.getmtime(filepath)
if os.path.exists(filepath[:-1]) :
mtime = max(mtime, os.path.getmtime(filepath[:-1]))
module.__dict__["__mtime_check__"] = time.time()
except OSError: pass
return mtime
def resolve_object(module, obj_str, arg=None, silent=0):
"""
This function traverses the objects separated by .
(period) to find the last one we're looking for:
From left to right, find objects, if it is
an unbound method of a class, instantiate the
class passing the request as single argument
'arg' is sometimes req, sometimes filter,
sometimes connection
"""
obj = module
for obj_str in obj_str.split('.'):
parent = obj
# don't throw attribute errors when silent
if silent and not hasattr(obj, obj_str):
return None
# this adds a little clarity if we have an attriute error
if obj == module and not hasattr(module, obj_str):
if hasattr(module, "__file__"):
s = "module '%s' contains no '%s'" % (module.__file__, obj_str)
raise AttributeError(s)
obj = getattr(obj, obj_str)
if inspect.isclass(parent):
# The parent of this object is a class,
# it needs to be instantiated
instance = parent(arg)
obj = getattr(instance, obj_str)
return obj
def build_cgi_env(req):
"""
Utility function that returns a dictionary of
CGI environment variables as described in
http://hoohoo.ncsa.uiuc.edu/cgi/env.html
"""
req.add_cgi_vars()
env = req.subprocess_env.copy()
if req.path_info and len(req.path_info) > 0:
env["SCRIPT_NAME"] = req.uri[:-len(req.path_info)]
else:
env["SCRIPT_NAME"] = req.uri
env["GATEWAY_INTERFACE"] = "Python-CGI/1.1"
# you may want to comment this out for better security
if "authorization" in req.headers_in:
env["HTTP_AUTHORIZATION"] = req.headers_in["authorization"]
return env
class NullIO(object):
""" Abstract IO
"""
def tell(self): return 0
def read(self, n = -1): return ""
def readline(self, length = None): return ""
def readlines(self): return []
def write(self, s): pass
def writelines(self, list):
self.write("".join(list))
def isatty(self): return 0
def flush(self): pass
def close(self): pass
def detach(self): pass
def seek(self, pos, mode = 0): pass
class CGIStdin(NullIO):
def __init__(self, req):
self.pos = 0
self.req = req
self.BLOCK = 65536 # 64K
# note that self.buf sometimes contains leftovers
# that were read, but not used when readline was used
self.buf = ""
def read(self, n = -1):
if n == 0:
return ""
if n == -1:
s = self.req.read(self.BLOCK)
while s:
self.buf = self.buf + s
self.pos = self.pos + len(s)
s = self.req.read(self.BLOCK)
result = self.buf
self.buf = ""
return result
else:
if self.buf:
s = self.buf[:n]
n = n - len(s)
else:
s = ""
s = s + self.req.read(n)
self.pos = self.pos + len(s)
return s
def readlines(self):
s = (self.buf + self.read()).split('\n')
return [s + '\n' for s in s]
def readline(self, n = -1):
if n == 0:
return ""
# fill up the buffer
self.buf = self.buf + self.req.read(self.BLOCK)
# look for \n in the buffer
i = self.buf.find('\n')
while i == -1: # if \n not found - read more
if (n != -1) and (len(self.buf) >= n): # we're past n
i = n - 1
break
x = len(self.buf)
self.buf = self.buf + self.req.read(self.BLOCK)
if len(self.buf) == x: # nothing read, eof
i = x - 1
break
i = self.buf.find('\n', x)
# carve out the piece, then shorten the buffer
result = self.buf[:i+1]
self.buf = self.buf[i+1:]
self.pos = self.pos + len(result)
return result
class CGIStdout(NullIO):
"""
Class that allows writing to the socket directly for CGI.
"""
def __init__(self, req):
self.pos = 0
self.req = req
self.headers_sent = 0
self.headers = ""
def write(self, s):
if not s: return
if not self.headers_sent:
self.headers = self.headers + s
# are headers over yet?
headers_over = 0
# first try RFC-compliant CRLF
ss = self.headers.split('\r\n\r\n', 1)
if len(ss) < 2:
# second try with \n\n
ss = self.headers.split('\n\n', 1)
if len(ss) >= 2:
headers_over = 1
else:
headers_over = 1
if headers_over:
# headers done, process them
ss[0] = ss[0].replace('\r\n', '\n')
lines = ss[0].split('\n')
for line in lines:
h, v = line.split(":", 1)
v = v.strip()
if h.lower() == "status":
status = int(v.split()[0])
self.req.status = status
elif h.lower() == "content-type":
self.req.content_type = v
self.req.headers_out[h] = v
else:
self.req.headers_out.add(h, v)
self.headers_sent = 1
# write the body if any at this point
self.req.write(ss[1])
else:
self.req.write(str(s))
self.pos = self.pos + len(s)
def tell(self): return self.pos
def setup_cgi(req):
"""
Replace sys.stdin and stdout with an objects that read/write to
the socket, as well as substitute the os.environ.
Returns (environ, stdin, stdout) which you must save and then use
with restore_nocgi().
"""
# save env
save_env = os.environ.copy()
si = sys.stdin
so = sys.stdout
os.environ.update(build_cgi_env(req))
sys.stdout = CGIStdout(req)
sys.stdin = CGIStdin(req)
sys.argv = [] # keeps cgi.py happy
return save_env, si, so
def restore_nocgi(sav_env, si, so):
""" see setup_cgi() """
osenv = os.environ
# restore env
for k in list(osenv.keys()):
del osenv[k]
for k in sav_env:
osenv[k] = sav_env[k]
sys.stdout = si
sys.stdin = so
interpreter = None
main_server = None
_callback = None
def register_cleanup(callback, data=None):
_apache.register_cleanup(interpreter, main_server, callback, data)
def init(name, server):
"""
This function is called by the server at startup time
"""
global interpreter
global main_server
interpreter = name
main_server = server
sys.argv = ["mod_python"]
global _callback
_callback = CallBack()
return _callback
## Some functions made public
make_table = _apache.table
log_error = _apache.log_error
table = _apache.table
config_tree = _apache.config_tree
server_root = _apache.server_root
mpm_query = _apache.mpm_query
exists_config_define = _apache.exists_config_define
stat = _apache.stat
## Some constants
HTTP_CONTINUE = 100
HTTP_SWITCHING_PROTOCOLS = 101
HTTP_PROCESSING = 102
HTTP_OK = 200
HTTP_CREATED = 201
HTTP_ACCEPTED = 202
HTTP_NON_AUTHORITATIVE = 203
HTTP_NO_CONTENT = 204
HTTP_RESET_CONTENT = 205
HTTP_PARTIAL_CONTENT = 206
HTTP_MULTI_STATUS = 207
HTTP_MULTIPLE_CHOICES = 300
HTTP_MOVED_PERMANENTLY = 301
HTTP_MOVED_TEMPORARILY = 302
HTTP_SEE_OTHER = 303
HTTP_NOT_MODIFIED = 304
HTTP_USE_PROXY = 305
HTTP_TEMPORARY_REDIRECT = 307
HTTP_BAD_REQUEST = 400
HTTP_UNAUTHORIZED = 401
HTTP_PAYMENT_REQUIRED = 402
HTTP_FORBIDDEN = 403
HTTP_NOT_FOUND = 404
HTTP_METHOD_NOT_ALLOWED = 405
HTTP_NOT_ACCEPTABLE = 406
HTTP_PROXY_AUTHENTICATION_REQUIRED= 407
HTTP_REQUEST_TIME_OUT = 408
HTTP_CONFLICT = 409
HTTP_GONE = 410
HTTP_LENGTH_REQUIRED = 411
HTTP_PRECONDITION_FAILED = 412
HTTP_REQUEST_ENTITY_TOO_LARGE = 413
HTTP_REQUEST_URI_TOO_LARGE = 414
HTTP_UNSUPPORTED_MEDIA_TYPE = 415
HTTP_RANGE_NOT_SATISFIABLE = 416
HTTP_EXPECTATION_FAILED = 417
HTTP_IM_A_TEAPOT = 418
HTTP_UNPROCESSABLE_ENTITY = 422
HTTP_LOCKED = 423
HTTP_FAILED_DEPENDENCY = 424
HTTP_UPGRADE_REQUIRED = 426
HTTP_INTERNAL_SERVER_ERROR = 500
HTTP_NOT_IMPLEMENTED = 501
HTTP_BAD_GATEWAY = 502
HTTP_SERVICE_UNAVAILABLE = 503
HTTP_GATEWAY_TIME_OUT = 504
HTTP_VERSION_NOT_SUPPORTED = 505
HTTP_VARIANT_ALSO_VARIES = 506
HTTP_INSUFFICIENT_STORAGE = 507
HTTP_NOT_EXTENDED = 510
# The APLOG constants in Apache are derived from syslog.h
# constants, so we do same here.
try:
import syslog
APLOG_EMERG = syslog.LOG_EMERG # system is unusable
APLOG_ALERT = syslog.LOG_ALERT # action must be taken immediately
APLOG_CRIT = syslog.LOG_CRIT # critical conditions
APLOG_ERR = syslog.LOG_ERR # error conditions
APLOG_WARNING = syslog.LOG_WARNING # warning conditions
APLOG_NOTICE = syslog.LOG_NOTICE # normal but significant condition
APLOG_INFO = syslog.LOG_INFO # informational
APLOG_DEBUG = syslog.LOG_DEBUG # debug-level messages
except ImportError:
APLOG_EMERG = 0
APLOG_ALERT = 1
APLOG_CRIT = 2
APLOG_ERR = 3
APLOG_WARNING = 4
APLOG_NOTICE = 5
APLOG_INFO = 6
APLOG_DEBUG = 7
APLOG_NOERRNO = 0 # DEPRECATED
OK = REQ_PROCEED = 0
DONE = -2
DECLINED = REQ_NOACTION = -1
_phase_handler_names = {}
for _phase in ["PythonPostReadRequestHandler",
"PythonTransHandler",
"PythonHeaderParserHandler",
"PythonInitHandler",
"PythonAccessHandler",
"PythonAuthenHandler",
"PythonAuthzHandler",
"PythonTypeHandler",
"PythonFixupHandler",
"PythonHandler",
"PythonLogHandler",
"PythonCleanupHandler"]:
_phase_handler_names[_phase] = _phase[len("python"):].lower()
_status_values = {
"postreadrequesthandler": [ DECLINED, OK ],
"transhandler": [ DECLINED ],
"headerparserhandler": [ DECLINED, OK ],
"inithandler": [ DECLINED, OK ],
"accesshandler": [ DECLINED, OK ],
"authenhandler": [ DECLINED ],
"authzhandler": [ DECLINED ],
"typehandler": [ DECLINED ],
"fixuphandler": [ DECLINED, OK ],
"handler": [ OK ],
"loghandler": [ DECLINED, OK ],
"cleanuphandler": [ OK ],
}
# constants for get_remote_host
REMOTE_HOST = 0
REMOTE_NAME = 1
REMOTE_NOLOOKUP = 2
REMOTE_DOUBLE_REV = 3
# legacy/mod_python things
REQ_ABORTED = HTTP_INTERNAL_SERVER_ERROR
REQ_EXIT = "REQ_EXIT"
SERVER_RETURN = _apache.SERVER_RETURN
# the req.finfo tuple
FINFO_MODE = 0
FINFO_INO = 1
FINFO_DEV = 2
FINFO_NLINK = 3
FINFO_UID = 4
FINFO_GID = 5
FINFO_SIZE = 6
FINFO_ATIME = 7
FINFO_MTIME = 8
FINFO_CTIME = 9
FINFO_FNAME = 10
FINFO_NAME = 11
FINFO_FILETYPE = 12
# the req.parsed_uri
URI_SCHEME = 0
URI_HOSTINFO = 1
URI_USER = 2
URI_PASSWORD = 3
URI_HOSTNAME = 4
URI_PORT = 5
URI_PATH = 6
URI_QUERY = 7
URI_FRAGMENT = 8
# for req.proxyreq
PROXYREQ_NONE = 0 # No proxy
PROXYREQ_PROXY = 1 # Standard proxy
PROXYREQ_REVERSE = 2 # Reverse proxy
PROXYREQ_RESPONSE = 3 # Origin response
# methods for req.allow_method()
M_GET = 0 # RFC 2616: HTTP
M_PUT = 1
M_POST = 2
M_DELETE = 3
M_CONNECT = 4
M_OPTIONS = 5
M_TRACE = 6 # RFC 2616: HTTP
M_PATCH = 7
M_PROPFIND = 8 # RFC 2518: WebDAV
M_PROPPATCH = 9
M_MKCOL = 10
M_COPY = 11
M_MOVE = 12
M_LOCK = 13
M_UNLOCK = 14 # RFC2518: WebDAV
M_VERSION_CONTROL = 15 # RFC3253: WebDAV Versioning
M_CHECKOUT = 16
M_UNCHECKOUT = 17
M_CHECKIN = 18
M_UPDATE = 19
M_LABEL = 20
M_REPORT = 21
M_MKWORKSPACE = 22
M_MKACTIVITY = 23
M_BASELINE_CONTROL = 24
M_MERGE = 25
M_INVALID = 26 # RFC3253: WebDAV Versioning
# for req.used_path_info
AP_REQ_ACCEPT_PATH_INFO = 0 # Accept request given path_info
AP_REQ_REJECT_PATH_INFO = 1 # Send 404 error if path_info was given
AP_REQ_DEFAULT_PATH_INFO = 2 # Module's choice for handling path_info
# for mpm_query
AP_MPMQ_NOT_SUPPORTED = 0 # This value specifies whether
# an MPM is capable of
# threading or forking.
AP_MPMQ_STATIC = 1 # This value specifies whether
# an MPM is using a static # of
# threads or daemons.
AP_MPMQ_DYNAMIC = 2 # This value specifies whether
# an MPM is using a dynamic # of
# threads or daemons.
AP_MPMQ_MAX_DAEMON_USED = 1 # Max # of daemons used so far
AP_MPMQ_IS_THREADED = 2 # MPM can do threading
AP_MPMQ_IS_FORKED = 3 # MPM can do forking
AP_MPMQ_HARD_LIMIT_DAEMONS = 4 # The compiled max # daemons
AP_MPMQ_HARD_LIMIT_THREADS = 5 # The compiled max # threads
AP_MPMQ_MAX_THREADS = 6 # # of threads/child by config
AP_MPMQ_MIN_SPARE_DAEMONS = 7 # Min # of spare daemons
AP_MPMQ_MIN_SPARE_THREADS = 8 # Min # of spare threads
AP_MPMQ_MAX_SPARE_DAEMONS = 9 # Max # of spare daemons
AP_MPMQ_MAX_SPARE_THREADS = 10 # Max # of spare threads
AP_MPMQ_MAX_REQUESTS_DAEMON= 11 # Max # of requests per daemon
AP_MPMQ_MAX_DAEMONS = 12 # Max # of daemons by config
# magic mime types
CGI_MAGIC_TYPE = "application/x-httpd-cgi"
INCLUDES_MAGIC_TYPE = "text/x-server-parsed-html"
INCLUDES_MAGIC_TYPE3 = "text/x-server-parsed-html3"
DIR_MAGIC_TYPE = "httpd/unix-directory"
# for req.read_body
REQUEST_NO_BODY = 0
REQUEST_CHUNKED_ERROR = 1
REQUEST_CHUNKED_DECHUNK = 2
# for req.connection.keepalive
AP_CONN_UNKNOWN = _apache.AP_CONN_UNKNOWN
AP_CONN_CLOSE = _apache.AP_CONN_CLOSE
AP_CONN_KEEPALIVE = _apache.AP_CONN_KEEPALIVE
# for req.finfo[apache.FINFO_FILETYPE]
APR_NOFILE = _apache.APR_NOFILE
APR_REG = _apache.APR_REG
APR_DIR = _apache.APR_DIR
APR_CHR = _apache.APR_CHR
APR_BLK = _apache.APR_BLK
APR_PIPE = _apache.APR_PIPE
APR_LNK = _apache.APR_LNK
APR_SOCK = _apache.APR_SOCK
APR_UNKFILE = _apache.APR_UNKFILE
# module magic
MODULE_MAGIC_NUMBER_MAJOR = _apache.MODULE_MAGIC_NUMBER_MAJOR
MODULE_MAGIC_NUMBER_MINOR = _apache.MODULE_MAGIC_NUMBER_MINOR
# for apache.stat()
APR_FINFO_LINK = 0x00000001 # Stat the link not the file itself if it is a link
APR_FINFO_MTIME = 0x00000010 # Modification Time
APR_FINFO_CTIME = 0x00000020 # Creation or inode-changed time
APR_FINFO_ATIME = 0x00000040 # Access Time
APR_FINFO_SIZE = 0x00000100 # Size of the file
APR_FINFO_CSIZE = 0x00000200 # Storage size consumed by the file
APR_FINFO_DEV = 0x00001000 # Device
APR_FINFO_INODE = 0x00002000 # Inode
APR_FINFO_NLINK = 0x00004000 # Number of links
APR_FINFO_TYPE = 0x00008000 # Type
APR_FINFO_USER = 0x00010000 # User
APR_FINFO_GROUP = 0x00020000 # Group
APR_FINFO_UPROT = 0x00100000 # User protection bits
APR_FINFO_GPROT = 0x00200000 # Group protection bits
APR_FINFO_WPROT = 0x00400000 # World protection bits
APR_FINFO_ICASE = 0x01000000 # if dev is case insensitive
APR_FINFO_NAME = 0x02000000 # ->name in proper case
APR_FINFO_MIN = 0x00008170 # type, mtime, ctime, atime, size
APR_FINFO_IDENT = 0x00003000 # dev and inode
APR_FINFO_OWNER = 0x00030000 # user and group
APR_FINFO_PROT = 0x00700000 # all protections
APR_FINFO_NORM = 0x0073b170 # an atomic unix apr_stat()
APR_FINFO_DIRENT = 0x02000000 # an atomic unix apr_dir_read()