# vim: set sw=4 expandtab : # # Copyright (C) 2000, 2001, 2013 Gregory Trubetskoy # Copyright (C) 2002, 2003, 2004, 2005, 2006, 2007 Apache Software Foundation # # Licensed under the Apache License, Version 2.0 (the "License"); you # may not use this file except in compliance with the License. You # may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or # implied. See the License for the specific language governing # permissions and limitations under the License. # # Originally developed by Gregory Trubetskoy. # import sys import traceback import time import os import pdb import stat import inspect import importlib import types import _apache try: from html import escape except: from cgi import escape try: import threading except: import dummy_threading as threading # Cache for values of PythonPath that have been seen already. _path_cache = {} _path_cache_lock = threading.Lock() _result_warning = """Handler has returned result or raised SERVER_RETURN exception with argument having non integer type. Type of value returned was %s, whereas expected """ + str(int) + "." class CallBack: """ A generic callback object. """ class HStack: """ The actual stack string lives in the request object so it can be manipulated by both apache.py and mod_python.c """ def __init__(self, req): self.req = req def pop(self): handlers = self.req.hstack.split() if not handlers: return None else: self.req.hstack = " ".join(handlers[1:]) return handlers[0] def ConnectionDispatch(self, conn): # config config, debug = conn.base_server.get_config(), False if "PythonDebug" in config: debug = config["PythonDebug"] == "1" try: handler = conn.hlist.handler # split module::handler l = handler.split('::', 1) module_name = l[0] if len(l) == 1: # no oject, provide default obj_str = "connectionhandler" else: obj_str = l[1] # evaluate pythonpath and set sys.path to # resulting value if not already done if "PythonPath" in config: _path_cache_lock.acquire() try: pathstring = config["PythonPath"] if pathstring not in _path_cache: newpath = eval(pathstring) _path_cache[pathstring] = None sys.path[:] = newpath finally: _path_cache_lock.release() # import module autoreload = True if "PythonAutoReload" in config: autoreload = config["PythonAutoReload"] == "1" module = import_module(module_name, autoreload=autoreload, log=debug) # find the object obj = resolve_object(module, obj_str, arg=conn, silent=0) # Only permit debugging using pdb if Apache has # actually been started in single process mode. pdb_debug = False if "PythonEnablePdb" in config: pdb_debug = config["PythonEnablePdb"] == "1" if pdb_debug and exists_config_define("ONE_PROCESS"): # Don't use pdb.runcall() as it results in # a bogus 'None' response when pdb session # is quit. With this code the exception # marking that the session has been quit is # propogated back up and it is obvious in # the error message what actually occurred. debugger = pdb.Pdb() debugger.reset() sys.settrace(debugger.trace_dispatch) try: result = obj(conn) finally: debugger.quitting = 1 sys.settrace(None) else: result = obj(conn) assert (result.__class__ is int), \ "ConnectionHandler '%s' returned invalid return code." % handler except: # Error (usually parsing) try: exc_type, exc_value, exc_traceback = sys.exc_info() result = self.ReportError(exc_type, exc_value, exc_traceback, srv=conn.base_server, phase="ConnectionHandler", hname=handler, debug=debug) finally: exc_traceback = None return result def FilterDispatch(self, fltr): req = fltr.req # config config, debug = req.get_config(), False if "PythonDebug" in config: debug = config["PythonDebug"] == "1" try: # split module::handler l = fltr.handler.split('::', 1) module_name = l[0] if len(l) == 1: # no oject, provide default if fltr.is_input: obj_str = "inputfilter" else: obj_str = "outputfilter" else: obj_str = l[1] # add the directory to pythonpath if # not there yet, or evaluate pythonpath # and set sys.path to resulting value # if not already done if "PythonPath" in config: _path_cache_lock.acquire() try: pathstring = config["PythonPath"] if pathstring not in _path_cache: newpath = eval(pathstring) _path_cache[pathstring] = None sys.path[:] = newpath finally: _path_cache_lock.release() else: if fltr.dir: _path_cache_lock.acquire() try: if fltr.dir not in sys.path: sys.path[:0] = [fltr.dir] finally: _path_cache_lock.release() # import module autoreload = True if "PythonAutoReload" in config: autoreload = config["PythonAutoReload"] == "1" module = import_module(module_name, autoreload=autoreload, log=debug) # find the object obj = resolve_object(module, obj_str, arg=fltr, silent=0) # Only permit debugging using pdb if Apache has # actually been started in single process mode. pdb_debug = False if "PythonEnablePdb" in config: pdb_debug = config["PythonEnablePdb"] == "1" if pdb_debug and exists_config_define("ONE_PROCESS"): # Don't use pdb.runcall() as it results in # a bogus 'None' response when pdb session # is quit. With this code the exception # marking that the session has been quit is # propogated back up and it is obvious in # the error message what actually occurred. debugger = pdb.Pdb() debugger.reset() sys.settrace(debugger.trace_dispatch) try: result = obj(fltr) finally: debugger.quitting = 1 sys.settrace(None) else: result = obj(fltr) # always flush the filter. without a FLUSH or EOS bucket, # the content is never written to the network. # XXX an alternative is to tell the user to flush() always if not fltr.closed: fltr.flush() except SERVER_RETURN as value: # SERVER_RETURN indicates a non-local abort from below # with value as (result, status) or (result, None) or result try: if len(value.args) == 2: (result, status) = value.args if status: req.status = status else: result = value.args[0] if result.__class__ is not int: s = "Value raised with SERVER_RETURN is invalid. It is a " s = s + "%s, but it must be a tuple or an int." % result.__class__ _apache.log_error(s, APLOG_ERR, req.server) return except: pass except: # Error (usually parsing) try: exc_type, exc_value, exc_traceback = sys.exc_info() fltr.disable() result = self.ReportError(exc_type, exc_value, exc_traceback, req=req, filter=fltr, phase=fltr.name, hname=fltr.handler, debug=debug) finally: exc_traceback = None return OK def HandlerDispatch(self, req): """ This is the handler dispatcher. """ # be cautious result = HTTP_INTERNAL_SERVER_ERROR # config config, debug = req.get_config(), False if "PythonDebug" in config: debug = config["PythonDebug"] == "1" default_obj_str = _phase_handler_names[req.phase] # Lookup expected status values that allow us to # continue when multiple handlers exist. expected = _status_values[default_obj_str] try: hlist = req.hlist while hlist.handler is not None: # split module::handler l = hlist.handler.split('::', 1) module_name = l[0] if len(l) == 1: # no object, provide default obj_str = default_obj_str else: obj_str = l[1] # add the directory to pythonpath if # not there yet, or evaluate pythonpath # and set sys.path to resulting value # if not already done if "PythonPath" in config: _path_cache_lock.acquire() try: pathstring = config["PythonPath"] if pathstring not in _path_cache: newpath = eval(pathstring) _path_cache[pathstring] = None sys.path[:] = newpath finally: _path_cache_lock.release() else: if not hlist.is_location: directory = hlist.directory if directory: _path_cache_lock.acquire() try: if directory not in sys.path: sys.path[:0] = [directory] finally: _path_cache_lock.release() # import module autoreload = True if "PythonAutoReload" in config: autoreload = config["PythonAutoReload"] == "1" module = import_module(module_name, autoreload=autoreload, log=debug) # find the object if '.' not in obj_str: # this is an optimization try: obj = module.__dict__[obj_str] except: if not hlist.silent: s = "module '%s' contains no '%s'" % (module.__file__, obj_str) raise AttributeError(s) else: obj = resolve_object(module, obj_str, arg=req, silent=hlist.silent) if not hlist.silent or obj is not None: try: # Only permit debugging using pdb if Apache has # actually been started in single process mode. pdb_debug = False if "PythonEnablePdb" in config: pdb_debug = config["PythonEnablePdb"] == "1" if pdb_debug and exists_config_define("ONE_PROCESS"): # Don't use pdb.runcall() as it results in # a bogus 'None' response when pdb session # is quit. With this code the exception # marking that the session has been quit is # propogated back up and it is obvious in # the error message what actually occurred. debugger = pdb.Pdb() debugger.reset() sys.settrace(debugger.trace_dispatch) try: result = obj(req) finally: debugger.quitting = 1 sys.settrace(None) else: result = obj(req) except SERVER_RETURN as value: # The SERVER_RETURN exception type when raised # otherwise indicates an abort from below with # value as (result, status) or (result, None) or # result. if len(value.args) == 2: (result, status) = value.args if status: req.status = status else: result = value.args[0] assert (result.__class__ is int), \ _result_warning % result.__class__ # stop cycling through handlers if result not in expected: break elif hlist.silent: # A missing handler when in silent mode will # only propagate DECLINED if it is the first # and only handler. if result == HTTP_INTERNAL_SERVER_ERROR: result = DECLINED hlist.next() except: # Error (usually parsing) try: exc_type, exc_value, exc_traceback = sys.exc_info() result = self.ReportError(exc_type, exc_value, exc_traceback, req=req, phase=req.phase, hname=hlist.handler, debug=debug) finally: exc_traceback = None return result def IncludeDispatch(self, fltr, tag, code): try: # config config, debug = fltr.req.get_config(), False if "PythonDebug" in config: debug = config["PythonDebug"] == "1" if not hasattr(fltr.req,"ssi_globals"): fltr.req.ssi_globals = {} fltr.req.ssi_globals["filter"] = fltr fltr.req.ssi_globals["__file__"] = fltr.req.filename code = code.replace('\r\n', '\n').rstrip() if tag == 'eval': result = eval(code, fltr.req.ssi_globals) if result is not None: fltr.write(str(result)) elif tag == 'exec': exec(code, fltr.req.ssi_globals) fltr.flush() except: try: exc_type, exc_value, exc_traceback = sys.exc_info() fltr.disable() result = self.ReportError(exc_type, exc_value, exc_traceback, req=fltr.req, filter=fltr, phase=fltr.name, hname=fltr.req.filename, debug=debug) finally: exc_traceback = None raise fltr.req.ssi_globals["filter"] = None return OK def ImportDispatch(self, name): # config config, debug = main_server.get_config(), False if "PythonDebug" in config: debug = config["PythonDebug"] == "1" # evaluate pythonpath and set sys.path to # resulting value if not already done if "PythonPath" in config: _path_cache_lock.acquire() try: pathstring = config["PythonPath"] if pathstring not in _path_cache: newpath = eval(pathstring) _path_cache[pathstring] = None sys.path[:] = newpath finally: _path_cache_lock.release() # split module::function l = name.split('::', 1) module_name = l[0] func_name = None if len(l) != 1: func_name = l[1] module = import_module(module_name, log=debug) if func_name: getattr(module, func_name)() def ReportError(self, etype, evalue, etb, req=None, filter=None, srv=None, phase="N/A", hname="N/A", debug=0): """ This function is only used when debugging is on. It sends the output similar to what you'd see when using Python interactively to the browser """ try: # try/finally try: # try/except if str(etype) == "exceptions.IOError" \ and str(evalue)[:5] == "Write": # if this is an IOError while writing to client, # it is probably better not to try to write to the cleint # even if debug is on. debug = 0 # write to log for e in traceback.format_exception(etype, evalue, etb): s = "%s %s: %s" % (phase, hname, e[:-1]) if req: req.log_error(s, APLOG_ERR) else: _apache.log_error(s, APLOG_ERR, srv) if not debug or not req: return HTTP_INTERNAL_SERVER_ERROR else: # write to client req.status = HTTP_INTERNAL_SERVER_ERROR req.content_type = 'text/html' s = '\n
\nMod_python error: "%s %s"\n\n' % (phase, hname)
for e in traceback.format_exception(etype, evalue, etb):
s = s + escape(e) + '\n'
s = s + "\n"
if filter:
filter.write(s)
filter.flush()
else:
req.write(s)
return DONE
except:
# last try
traceback.print_exc()
sys.stderr.flush()
finally:
# erase the traceback
etb = None
# we do not return anything
def import_module(module_name, autoreload=1, log=0, path=None):
"""
Get the module to handle the request. If
autoreload is on, then the module will be reloaded
if it has changed since the last import.
"""
# nlehuen: this is a big lock, we'll have to refine it later to get better performance.
# For now, we'll concentrate on thread-safety.
# (Re)import
if module_name in sys.modules:
# The module has been imported already
module = sys.modules[module_name]
oldmtime, mtime = 0, 0
if autoreload:
# but is it in the path?
try:
file = module.__dict__["__file__"]
except KeyError:
file = None
# the "and not" part of this condition is to prevent execution
# of arbitrary already imported modules, such as os. The
# reason we use startswith as opposed to exact match is that
# modules inside packages are actually in subdirectories.
if not file or (path and not list(filter(file.startswith, path))):
# there is a script by this name already imported, but it's in
# a different directory, therefore it's a different script
mtime, oldmtime = 0, -1 # trigger import
else:
try:
last_check = module.__dict__["__mtime_check__"]
except KeyError:
last_check = 0
if (time.time() - last_check) > 1:
oldmtime = module.__dict__.get("__mtime__", 0)
mtime = module_mtime(module)
else:
pass
else:
mtime, oldmtime = 0, -1
if mtime != oldmtime:
# Import the module
if log:
if path:
s = "mod_python: (Re)importing module '%s' with path set to '%s'" % (module_name, path)
else:
s = "mod_python: (Re)importing module '%s'" % module_name
_apache.log_error(s, APLOG_NOTICE)
parent = None
module = importlib.import_module(module_name)
if mtime == 0:
mtime = module_mtime(module)
module.__mtime__ = mtime
return module
def module_mtime(module):
"""Get modification time of module"""
mtime = 0
if "__file__" in module.__dict__:
filepath = module.__file__
try:
# this try/except block is a workaround for a Python bug in
# 2.0, 2.1 and 2.1.1. See
# http://sourceforge.net/tracker/?group_id=5470&atid=105470&func=detail&aid=422004
if os.path.exists(filepath):
mtime = os.path.getmtime(filepath)
if os.path.exists(filepath[:-1]) :
mtime = max(mtime, os.path.getmtime(filepath[:-1]))
module.__dict__["__mtime_check__"] = time.time()
except OSError: pass
return mtime
def resolve_object(module, obj_str, arg=None, silent=0):
"""
This function traverses the objects separated by .
(period) to find the last one we're looking for:
From left to right, find objects, if it is
an unbound method of a class, instantiate the
class passing the request as single argument
'arg' is sometimes req, sometimes filter,
sometimes connection
"""
obj = module
for obj_str in obj_str.split('.'):
parent = obj
# don't throw attribute errors when silent
if silent and not hasattr(obj, obj_str):
return None
# this adds a little clarity if we have an attriute error
if obj == module and not hasattr(module, obj_str):
if hasattr(module, "__file__"):
s = "module '%s' contains no '%s'" % (module.__file__, obj_str)
raise AttributeError(s)
obj = getattr(obj, obj_str)
if inspect.isclass(parent):
# The parent of this object is a class,
# it needs to be instantiated
instance = parent(arg)
obj = getattr(instance, obj_str)
return obj
def build_cgi_env(req):
"""
Utility function that returns a dictionary of
CGI environment variables as described in
http://hoohoo.ncsa.uiuc.edu/cgi/env.html
"""
req.add_cgi_vars()
env = req.subprocess_env.copy()
if req.path_info and len(req.path_info) > 0:
env["SCRIPT_NAME"] = req.uri[:-len(req.path_info)]
else:
env["SCRIPT_NAME"] = req.uri
env["GATEWAY_INTERFACE"] = "Python-CGI/1.1"
# you may want to comment this out for better security
if "authorization" in req.headers_in:
env["HTTP_AUTHORIZATION"] = req.headers_in["authorization"]
return env
class NullIO(object):
""" Abstract IO
"""
def tell(self): return 0
def read(self, n = -1): return ""
def readline(self, length = None): return ""
def readlines(self): return []
def write(self, s): pass
def writelines(self, list):
self.write("".join(list))
def isatty(self): return 0
def flush(self): pass
def close(self): pass
def detach(self): pass
def seek(self, pos, mode = 0): pass
class CGIStdin(NullIO):
def __init__(self, req):
self.pos = 0
self.req = req
self.BLOCK = 65536 # 64K
# note that self.buf sometimes contains leftovers
# that were read, but not used when readline was used
self.buf = ""
def read(self, n = -1):
if n == 0:
return ""
if n == -1:
s = self.req.read(self.BLOCK)
while s:
self.buf = self.buf + s
self.pos = self.pos + len(s)
s = self.req.read(self.BLOCK)
result = self.buf
self.buf = ""
return result
else:
if self.buf:
s = self.buf[:n]
n = n - len(s)
else:
s = ""
s = s + self.req.read(n)
self.pos = self.pos + len(s)
return s
def readlines(self):
s = (self.buf + self.read()).split('\n')
return [s + '\n' for s in s]
def readline(self, n = -1):
if n == 0:
return ""
# fill up the buffer
self.buf = self.buf + self.req.read(self.BLOCK)
# look for \n in the buffer
i = self.buf.find('\n')
while i == -1: # if \n not found - read more
if (n != -1) and (len(self.buf) >= n): # we're past n
i = n - 1
break
x = len(self.buf)
self.buf = self.buf + self.req.read(self.BLOCK)
if len(self.buf) == x: # nothing read, eof
i = x - 1
break
i = self.buf.find('\n', x)
# carve out the piece, then shorten the buffer
result = self.buf[:i+1]
self.buf = self.buf[i+1:]
self.pos = self.pos + len(result)
return result
class CGIStdout(NullIO):
"""
Class that allows writing to the socket directly for CGI.
"""
def __init__(self, req):
self.pos = 0
self.req = req
self.headers_sent = 0
self.headers = ""
def write(self, s):
if not s: return
if not self.headers_sent:
self.headers = self.headers + s
# are headers over yet?
headers_over = 0
# first try RFC-compliant CRLF
ss = self.headers.split('\r\n\r\n', 1)
if len(ss) < 2:
# second try with \n\n
ss = self.headers.split('\n\n', 1)
if len(ss) >= 2:
headers_over = 1
else:
headers_over = 1
if headers_over:
# headers done, process them
ss[0] = ss[0].replace('\r\n', '\n')
lines = ss[0].split('\n')
for line in lines:
h, v = line.split(":", 1)
v = v.strip()
if h.lower() == "status":
status = int(v.split()[0])
self.req.status = status
elif h.lower() == "content-type":
self.req.content_type = v
self.req.headers_out[h] = v
else:
self.req.headers_out.add(h, v)
self.headers_sent = 1
# write the body if any at this point
self.req.write(ss[1])
else:
self.req.write(str(s))
self.pos = self.pos + len(s)
def tell(self): return self.pos
def setup_cgi(req):
"""
Replace sys.stdin and stdout with an objects that read/write to
the socket, as well as substitute the os.environ.
Returns (environ, stdin, stdout) which you must save and then use
with restore_nocgi().
"""
# save env
save_env = os.environ.copy()
si = sys.stdin
so = sys.stdout
os.environ.update(build_cgi_env(req))
sys.stdout = CGIStdout(req)
sys.stdin = CGIStdin(req)
sys.argv = [] # keeps cgi.py happy
return save_env, si, so
def restore_nocgi(sav_env, si, so):
""" see setup_cgi() """
osenv = os.environ
# restore env
for k in list(osenv.keys()):
del osenv[k]
for k in sav_env:
osenv[k] = sav_env[k]
sys.stdout = si
sys.stdin = so
interpreter = None
main_server = None
_callback = None
def register_cleanup(callback, data=None):
_apache.register_cleanup(interpreter, main_server, callback, data)
def init(name, server):
"""
This function is called by the server at startup time
"""
global interpreter
global main_server
interpreter = name
main_server = server
sys.argv = ["mod_python"]
global _callback
_callback = CallBack()
return _callback
## Some functions made public
make_table = _apache.table
log_error = _apache.log_error
table = _apache.table
config_tree = _apache.config_tree
server_root = _apache.server_root
mpm_query = _apache.mpm_query
exists_config_define = _apache.exists_config_define
stat = _apache.stat
## Some constants
HTTP_CONTINUE = 100
HTTP_SWITCHING_PROTOCOLS = 101
HTTP_PROCESSING = 102
HTTP_OK = 200
HTTP_CREATED = 201
HTTP_ACCEPTED = 202
HTTP_NON_AUTHORITATIVE = 203
HTTP_NO_CONTENT = 204
HTTP_RESET_CONTENT = 205
HTTP_PARTIAL_CONTENT = 206
HTTP_MULTI_STATUS = 207
HTTP_MULTIPLE_CHOICES = 300
HTTP_MOVED_PERMANENTLY = 301
HTTP_MOVED_TEMPORARILY = 302
HTTP_SEE_OTHER = 303
HTTP_NOT_MODIFIED = 304
HTTP_USE_PROXY = 305
HTTP_TEMPORARY_REDIRECT = 307
HTTP_BAD_REQUEST = 400
HTTP_UNAUTHORIZED = 401
HTTP_PAYMENT_REQUIRED = 402
HTTP_FORBIDDEN = 403
HTTP_NOT_FOUND = 404
HTTP_METHOD_NOT_ALLOWED = 405
HTTP_NOT_ACCEPTABLE = 406
HTTP_PROXY_AUTHENTICATION_REQUIRED= 407
HTTP_REQUEST_TIME_OUT = 408
HTTP_CONFLICT = 409
HTTP_GONE = 410
HTTP_LENGTH_REQUIRED = 411
HTTP_PRECONDITION_FAILED = 412
HTTP_REQUEST_ENTITY_TOO_LARGE = 413
HTTP_REQUEST_URI_TOO_LARGE = 414
HTTP_UNSUPPORTED_MEDIA_TYPE = 415
HTTP_RANGE_NOT_SATISFIABLE = 416
HTTP_EXPECTATION_FAILED = 417
HTTP_IM_A_TEAPOT = 418
HTTP_UNPROCESSABLE_ENTITY = 422
HTTP_LOCKED = 423
HTTP_FAILED_DEPENDENCY = 424
HTTP_UPGRADE_REQUIRED = 426
HTTP_INTERNAL_SERVER_ERROR = 500
HTTP_NOT_IMPLEMENTED = 501
HTTP_BAD_GATEWAY = 502
HTTP_SERVICE_UNAVAILABLE = 503
HTTP_GATEWAY_TIME_OUT = 504
HTTP_VERSION_NOT_SUPPORTED = 505
HTTP_VARIANT_ALSO_VARIES = 506
HTTP_INSUFFICIENT_STORAGE = 507
HTTP_NOT_EXTENDED = 510
# The APLOG constants in Apache are derived from syslog.h
# constants, so we do same here.
try:
import syslog
APLOG_EMERG = syslog.LOG_EMERG # system is unusable
APLOG_ALERT = syslog.LOG_ALERT # action must be taken immediately
APLOG_CRIT = syslog.LOG_CRIT # critical conditions
APLOG_ERR = syslog.LOG_ERR # error conditions
APLOG_WARNING = syslog.LOG_WARNING # warning conditions
APLOG_NOTICE = syslog.LOG_NOTICE # normal but significant condition
APLOG_INFO = syslog.LOG_INFO # informational
APLOG_DEBUG = syslog.LOG_DEBUG # debug-level messages
except ImportError:
APLOG_EMERG = 0
APLOG_ALERT = 1
APLOG_CRIT = 2
APLOG_ERR = 3
APLOG_WARNING = 4
APLOG_NOTICE = 5
APLOG_INFO = 6
APLOG_DEBUG = 7
APLOG_NOERRNO = 0 # DEPRECATED
OK = REQ_PROCEED = 0
DONE = -2
DECLINED = REQ_NOACTION = -1
_phase_handler_names = {}
for _phase in ["PythonPostReadRequestHandler",
"PythonTransHandler",
"PythonHeaderParserHandler",
"PythonInitHandler",
"PythonAccessHandler",
"PythonAuthenHandler",
"PythonAuthzHandler",
"PythonTypeHandler",
"PythonFixupHandler",
"PythonHandler",
"PythonLogHandler",
"PythonCleanupHandler"]:
_phase_handler_names[_phase] = _phase[len("python"):].lower()
_status_values = {
"postreadrequesthandler": [ DECLINED, OK ],
"transhandler": [ DECLINED ],
"headerparserhandler": [ DECLINED, OK ],
"inithandler": [ DECLINED, OK ],
"accesshandler": [ DECLINED, OK ],
"authenhandler": [ DECLINED ],
"authzhandler": [ DECLINED ],
"typehandler": [ DECLINED ],
"fixuphandler": [ DECLINED, OK ],
"handler": [ OK ],
"loghandler": [ DECLINED, OK ],
"cleanuphandler": [ OK ],
}
# constants for get_remote_host
REMOTE_HOST = 0
REMOTE_NAME = 1
REMOTE_NOLOOKUP = 2
REMOTE_DOUBLE_REV = 3
# legacy/mod_python things
REQ_ABORTED = HTTP_INTERNAL_SERVER_ERROR
REQ_EXIT = "REQ_EXIT"
SERVER_RETURN = _apache.SERVER_RETURN
# the req.finfo tuple
FINFO_MODE = 0
FINFO_INO = 1
FINFO_DEV = 2
FINFO_NLINK = 3
FINFO_UID = 4
FINFO_GID = 5
FINFO_SIZE = 6
FINFO_ATIME = 7
FINFO_MTIME = 8
FINFO_CTIME = 9
FINFO_FNAME = 10
FINFO_NAME = 11
FINFO_FILETYPE = 12
# the req.parsed_uri
URI_SCHEME = 0
URI_HOSTINFO = 1
URI_USER = 2
URI_PASSWORD = 3
URI_HOSTNAME = 4
URI_PORT = 5
URI_PATH = 6
URI_QUERY = 7
URI_FRAGMENT = 8
# for req.proxyreq
PROXYREQ_NONE = 0 # No proxy
PROXYREQ_PROXY = 1 # Standard proxy
PROXYREQ_REVERSE = 2 # Reverse proxy
PROXYREQ_RESPONSE = 3 # Origin response
# methods for req.allow_method()
M_GET = 0 # RFC 2616: HTTP
M_PUT = 1
M_POST = 2
M_DELETE = 3
M_CONNECT = 4
M_OPTIONS = 5
M_TRACE = 6 # RFC 2616: HTTP
M_PATCH = 7
M_PROPFIND = 8 # RFC 2518: WebDAV
M_PROPPATCH = 9
M_MKCOL = 10
M_COPY = 11
M_MOVE = 12
M_LOCK = 13
M_UNLOCK = 14 # RFC2518: WebDAV
M_VERSION_CONTROL = 15 # RFC3253: WebDAV Versioning
M_CHECKOUT = 16
M_UNCHECKOUT = 17
M_CHECKIN = 18
M_UPDATE = 19
M_LABEL = 20
M_REPORT = 21
M_MKWORKSPACE = 22
M_MKACTIVITY = 23
M_BASELINE_CONTROL = 24
M_MERGE = 25
M_INVALID = 26 # RFC3253: WebDAV Versioning
# for req.used_path_info
AP_REQ_ACCEPT_PATH_INFO = 0 # Accept request given path_info
AP_REQ_REJECT_PATH_INFO = 1 # Send 404 error if path_info was given
AP_REQ_DEFAULT_PATH_INFO = 2 # Module's choice for handling path_info
# for mpm_query
AP_MPMQ_NOT_SUPPORTED = 0 # This value specifies whether
# an MPM is capable of
# threading or forking.
AP_MPMQ_STATIC = 1 # This value specifies whether
# an MPM is using a static # of
# threads or daemons.
AP_MPMQ_DYNAMIC = 2 # This value specifies whether
# an MPM is using a dynamic # of
# threads or daemons.
AP_MPMQ_MAX_DAEMON_USED = 1 # Max # of daemons used so far
AP_MPMQ_IS_THREADED = 2 # MPM can do threading
AP_MPMQ_IS_FORKED = 3 # MPM can do forking
AP_MPMQ_HARD_LIMIT_DAEMONS = 4 # The compiled max # daemons
AP_MPMQ_HARD_LIMIT_THREADS = 5 # The compiled max # threads
AP_MPMQ_MAX_THREADS = 6 # # of threads/child by config
AP_MPMQ_MIN_SPARE_DAEMONS = 7 # Min # of spare daemons
AP_MPMQ_MIN_SPARE_THREADS = 8 # Min # of spare threads
AP_MPMQ_MAX_SPARE_DAEMONS = 9 # Max # of spare daemons
AP_MPMQ_MAX_SPARE_THREADS = 10 # Max # of spare threads
AP_MPMQ_MAX_REQUESTS_DAEMON= 11 # Max # of requests per daemon
AP_MPMQ_MAX_DAEMONS = 12 # Max # of daemons by config
# magic mime types
CGI_MAGIC_TYPE = "application/x-httpd-cgi"
INCLUDES_MAGIC_TYPE = "text/x-server-parsed-html"
INCLUDES_MAGIC_TYPE3 = "text/x-server-parsed-html3"
DIR_MAGIC_TYPE = "httpd/unix-directory"
# for req.read_body
REQUEST_NO_BODY = 0
REQUEST_CHUNKED_ERROR = 1
REQUEST_CHUNKED_DECHUNK = 2
# for req.connection.keepalive
AP_CONN_UNKNOWN = _apache.AP_CONN_UNKNOWN
AP_CONN_CLOSE = _apache.AP_CONN_CLOSE
AP_CONN_KEEPALIVE = _apache.AP_CONN_KEEPALIVE
# for req.finfo[apache.FINFO_FILETYPE]
APR_NOFILE = _apache.APR_NOFILE
APR_REG = _apache.APR_REG
APR_DIR = _apache.APR_DIR
APR_CHR = _apache.APR_CHR
APR_BLK = _apache.APR_BLK
APR_PIPE = _apache.APR_PIPE
APR_LNK = _apache.APR_LNK
APR_SOCK = _apache.APR_SOCK
APR_UNKFILE = _apache.APR_UNKFILE
# module magic
MODULE_MAGIC_NUMBER_MAJOR = _apache.MODULE_MAGIC_NUMBER_MAJOR
MODULE_MAGIC_NUMBER_MINOR = _apache.MODULE_MAGIC_NUMBER_MINOR
# for apache.stat()
APR_FINFO_LINK = 0x00000001 # Stat the link not the file itself if it is a link
APR_FINFO_MTIME = 0x00000010 # Modification Time
APR_FINFO_CTIME = 0x00000020 # Creation or inode-changed time
APR_FINFO_ATIME = 0x00000040 # Access Time
APR_FINFO_SIZE = 0x00000100 # Size of the file
APR_FINFO_CSIZE = 0x00000200 # Storage size consumed by the file
APR_FINFO_DEV = 0x00001000 # Device
APR_FINFO_INODE = 0x00002000 # Inode
APR_FINFO_NLINK = 0x00004000 # Number of links
APR_FINFO_TYPE = 0x00008000 # Type
APR_FINFO_USER = 0x00010000 # User
APR_FINFO_GROUP = 0x00020000 # Group
APR_FINFO_UPROT = 0x00100000 # User protection bits
APR_FINFO_GPROT = 0x00200000 # Group protection bits
APR_FINFO_WPROT = 0x00400000 # World protection bits
APR_FINFO_ICASE = 0x01000000 # if dev is case insensitive
APR_FINFO_NAME = 0x02000000 # ->name in proper case
APR_FINFO_MIN = 0x00008170 # type, mtime, ctime, atime, size
APR_FINFO_IDENT = 0x00003000 # dev and inode
APR_FINFO_OWNER = 0x00030000 # user and group
APR_FINFO_PROT = 0x00700000 # all protections
APR_FINFO_NORM = 0x0073b170 # an atomic unix apr_stat()
APR_FINFO_DIRENT = 0x02000000 # an atomic unix apr_dir_read()