X-Git-Url: http://git.inspyration.org/?a=blobdiff_plain;f=openerp%2Fnetsvc.py;h=bfdf7d39352d437ca2af87740b218b407e96fe41;hb=fcd06145dda957a83e687fd033ff294c407da5f4;hp=7418aee2661cadc0432c4b45138e1172c1e142db;hpb=aca0720691db7a6c2bff007d4547374baefd771c;p=odoo%2Fodoo.git diff --git a/openerp/netsvc.py b/openerp/netsvc.py index 7418aee..bfdf7d3 100644 --- a/openerp/netsvc.py +++ b/openerp/netsvc.py @@ -1,9 +1,8 @@ -#!/usr/bin/env python # -*- coding: utf-8 -*- ############################################################################## # # OpenERP, Open Source Management Solution -# Copyright (C) 2004-2011 OpenERP SA () +# Copyright (C) 2004-2014 OpenERP SA () # # This program is free software: you can redistribute it and/or modify # it under the terms of the GNU Affero General Public License as @@ -20,102 +19,82 @@ # ############################################################################## -import errno import logging import logging.handlers import os -import platform +import pprint import release -import socket import sys import threading -import time -import types -from pprint import pformat -# TODO modules that import netsvc only for things from loglevels must be changed to use loglevels. -from loglevels import * -import tools +import psycopg2 + import openerp +import sql_db +import tools _logger = logging.getLogger(__name__) -def close_socket(sock): - """ Closes a socket instance cleanly +def log(logger, level, prefix, msg, depth=None): + indent='' + indent_after=' '*len(prefix) + for line in (prefix + pprint.pformat(msg, depth=depth)).split('\n'): + logger.log(level, indent+line) + indent=indent_after - :param sock: the network socket to close - :type sock: socket.socket +def LocalService(name): """ - try: - sock.shutdown(socket.SHUT_RDWR) - except socket.error, e: - # On OSX, socket shutdowns both sides if any side closes it - # causing an error 57 'Socket is not connected' on shutdown - # of the other side (or something), see - # http://bugs.python.org/issue4397 - # note: stdlib fixed test, not behavior - if e.errno != errno.ENOTCONN or platform.system() != 'Darwin': - raise - sock.close() - - -#.apidoc title: Common Services: netsvc -#.apidoc module-mods: member-order: bysource - -def abort_response(dummy_1, description, dummy_2, details): - # TODO Replace except_{osv,orm} with these directly. - raise openerp.osv.osv.except_osv(description, details) - -class Service(object): - """ Base class for *Local* services - - Functionality here is trusted, no authentication. + The openerp.netsvc.LocalService() function is deprecated. It still works + in two cases: workflows and reports. For workflows, instead of using + LocalService('workflow'), openerp.workflow should be used (better yet, + methods on openerp.osv.orm.Model should be used). For reports, + openerp.report.render_report() should be used (methods on the Model should + be provided too in the future). """ - _services = {} - def __init__(self, name): - Service._services[name] = self - self.__name = name + assert openerp.conf.deprecation.allow_local_service + _logger.warning("LocalService() is deprecated since march 2013 (it was called with '%s')." % name) - @classmethod - def exists(cls, name): - return name in cls._services - - @classmethod - def remove(cls, name): - if cls.exists(name): - cls._services.pop(name) - -def LocalService(name): - # Special case for addons support, will be removed in a few days when addons - # are updated to directly use openerp.osv.osv.service. - if name == 'object_proxy': - return openerp.osv.osv.service + if name == 'workflow': + return openerp.workflow - return Service._services[name] + if name.startswith('report.'): + report = openerp.report.interface.report_int._reports.get(name) + if report: + return report + else: + dbname = getattr(threading.currentThread(), 'dbname', None) + if dbname: + registry = openerp.modules.registry.RegistryManager.get(dbname) + with registry.cursor() as cr: + return registry['ir.actions.report.xml']._lookup_report(cr, name[len('report.'):]) -class ExportService(object): - """ Proxy for exported services. +path_prefix = os.path.realpath(os.path.dirname(os.path.dirname(__file__))) - Note that this class has no direct proxy, capable of calling - eservice.method(). Rather, the proxy should call - dispatch(method, params) +class PostgreSQLHandler(logging.Handler): + """ PostgreSQL Loggin Handler will store logs in the database, by default + the current database, can be set using --log-db=DBNAME """ - - _services = {} - - def __init__(self, name): - ExportService._services[name] = self - self.__name = name - _logger.debug("Registered an exported service: %s" % name) - - @classmethod - def getService(cls,name): - return cls._services[name] - - # Dispatch a RPC call w.r.t. the method name. The dispatching - # w.r.t. the service (this class) is done by OpenERPDispatcher. - def dispatch(self, method, params): - raise Exception("stub dispatch at %s" % self.__name) + def emit(self, record): + ct = threading.current_thread() + ct_db = getattr(ct, 'dbname', None) + dbname = tools.config['log_db'] or ct_db + if not dbname: + return + with tools.ignore(Exception), tools.mute_logger('openerp.sql_db'), sql_db.db_connect(dbname, allow_uri=True).cursor() as cr: + msg = tools.ustr(record.msg) + if record.args: + msg = msg % record.args + traceback = getattr(record, 'exc_text', '') + if traceback: + msg = "%s\n%s" % (msg, traceback) + # we do not use record.levelname because it may have been changed by ColoredFormatter. + levelname = logging.getLevelName(record.levelno) + + val = ('server', ct_db, record.name, levelname, msg, record.pathname[len(path_prefix)+1:], record.lineno, record.funcName) + cr.execute(""" + INSERT INTO ir_logging(create_date, type, dbname, name, level, message, path, line, func) + VALUES (NOW() at time zone 'UTC', %s, %s, %s, %s, %s, %s, %s, %s) + """, val) BLACK, RED, GREEN, YELLOW, BLUE, MAGENTA, CYAN, WHITE, _NOTHING, DEFAULT = range(10) #The background is set with 40 plus the number of the color, and the foreground with 30 @@ -125,12 +104,8 @@ COLOR_SEQ = "\033[1;%dm" BOLD_SEQ = "\033[1m" COLOR_PATTERN = "%s%s%%s%s" % (COLOR_SEQ, COLOR_SEQ, RESET_SEQ) LEVEL_COLOR_MAPPING = { - logging.DEBUG_SQL: (WHITE, MAGENTA), - logging.DEBUG_RPC: (BLUE, WHITE), - logging.DEBUG_RPC_ANSWER: (BLUE, WHITE), logging.DEBUG: (BLUE, DEFAULT), logging.INFO: (GREEN, DEFAULT), - logging.TEST: (WHITE, BLUE), logging.WARNING: (YELLOW, DEFAULT), logging.ERROR: (RED, DEFAULT), logging.CRITICAL: (WHITE, RED), @@ -144,17 +119,23 @@ class DBFormatter(logging.Formatter): class ColoredFormatter(DBFormatter): def format(self, record): - fg_color, bg_color = LEVEL_COLOR_MAPPING[record.levelno] + fg_color, bg_color = LEVEL_COLOR_MAPPING.get(record.levelno, (GREEN, DEFAULT)) record.levelname = COLOR_PATTERN % (30 + fg_color, 40 + bg_color, record.levelname) return DBFormatter.format(self, record) - +_logger_init = False def init_logger(): + global _logger_init + if _logger_init: + return + _logger_init = True + + logging.addLevelName(25, "INFO") + from tools.translate import resetlocale resetlocale() # create a format for log messages and dates - format = '[%(asctime)s][%(dbname)s] %(levelname)s:%(name)s:%(message)s' format = '%(asctime)s %(pid)s %(levelname)s %(dbname)s %(name)s: %(message)s' if tools.config['syslog']: @@ -162,7 +143,7 @@ def init_logger(): if os.name == 'nt': handler = logging.handlers.NTEventLogHandler("%s %s" % (release.description, release.version)) else: - handler = logging.handlers.SysLogHandler('/dev/log') + handler = logging.handlers.SysLogHandler() format = '%s %s' % (release.description, release.version) \ + ':%(dbname)s:%(levelname)s:%(name)s:%(message)s' @@ -170,15 +151,16 @@ def init_logger(): # LogFile Handler logf = tools.config['logfile'] try: + # We check we have the right location for the log files dirname = os.path.dirname(logf) if dirname and not os.path.isdir(dirname): os.makedirs(dirname) if tools.config['logrotate'] is not False: - handler = logging.handlers.TimedRotatingFileHandler(logf,'D',1,30) + handler = logging.handlers.TimedRotatingFileHandler(filename=logf, when='D', interval=1, backupCount=30) elif os.name == 'posix': handler = logging.handlers.WatchedFileHandler(logf) else: - handler = logging.handlers.FileHandler(logf) + handler = logging.FileHandler(logf) except Exception: sys.stderr.write("ERROR: couldn't create the logfile directory. Logging to the standard output.\n") handler = logging.StreamHandler(sys.stdout) @@ -186,176 +168,58 @@ def init_logger(): # Normal Handler on standard output handler = logging.StreamHandler(sys.stdout) - if isinstance(handler, logging.StreamHandler) and os.isatty(handler.stream.fileno()): + # Check that handler.stream has a fileno() method: when running OpenERP + # behind Apache with mod_wsgi, handler.stream will have type mod_wsgi.Log, + # which has no fileno() method. (mod_wsgi.Log is what is being bound to + # sys.stderr when the logging.StreamHandler is being constructed above.) + def is_a_tty(stream): + return hasattr(stream, 'fileno') and os.isatty(stream.fileno()) + + if isinstance(handler, logging.StreamHandler) and is_a_tty(handler.stream): formatter = ColoredFormatter(format) else: formatter = DBFormatter(format) handler.setFormatter(formatter) - # Configure handlers - logconfig = tools.config['log_handler'] - for i in logconfig: - prefix, level = i.split(':') - level = getattr(logging, level, logging.INFO) - logger = logging.getLogger(prefix) - logger.handlers = [] - logger.setLevel(level) - logger.addHandler(handler) - if prefix != '': - logger.propagate = False - + logging.getLogger().addHandler(handler) -# A alternative logging scheme for automated runs of the -# server intended to test it. -def init_alternative_logger(): - class H(logging.Handler): - def emit(self, record): - if record.levelno > 20: - print record.levelno, record.pathname, record.msg - handler = H() - # Add the handler to the 'openerp' logger. - logger = logging.getLogger('openerp') - logger.addHandler(handler) - logger.setLevel(logging.ERROR) + if tools.config['log_db']: + postgresqlHandler = PostgreSQLHandler() + postgresqlHandler.setLevel(25) + logging.getLogger().addHandler(postgresqlHandler) -class Server: - """ Generic interface for all servers with an event loop etc. - Override this to impement http, net-rpc etc. servers. + # Configure loggers levels + pseudo_config = PSEUDOCONFIG_MAPPER.get(tools.config['log_level'], []) - Servers here must have threaded behaviour. start() must not block, - there is no run(). - """ - __is_started = False - __servers = [] - __starter_threads = [] - - # we don't want blocking server calls (think select()) to - # wait forever and possibly prevent exiting the process, - # but instead we want a form of polling/busy_wait pattern, where - # _server_timeout should be used as the default timeout for - # all I/O blocking operations - _busywait_timeout = 0.5 - - def __init__(self): - Server.__servers.append(self) - if Server.__is_started: - # raise Exception('All instances of servers must be inited before the startAll()') - # Since the startAll() won't be called again, allow this server to - # init and then start it after 1sec (hopefully). Register that - # timer thread in a list, so that we can abort the start if quitAll - # is called in the meantime - t = threading.Timer(1.0, self._late_start) - t.name = 'Late start timer for %s' % str(self.__class__) - Server.__starter_threads.append(t) - t.start() - - def start(self): - _logger.debug("called stub Server.start") - - def _late_start(self): - self.start() - for thr in Server.__starter_threads: - if thr.finished.is_set(): - Server.__starter_threads.remove(thr) - - def stop(self): - _logger.debug("called stub Server.stop") - - def stats(self): - """ This function should return statistics about the server """ - return "%s: No statistics" % str(self.__class__) - - @classmethod - def startAll(cls): - if cls.__is_started: - return - _logger.info("Starting %d services" % len(cls.__servers)) - for srv in cls.__servers: - srv.start() - cls.__is_started = True - - @classmethod - def quitAll(cls): - if not cls.__is_started: - return - _logger.info("Stopping %d services" % len(cls.__servers)) - for thr in cls.__starter_threads: - if not thr.finished.is_set(): - thr.cancel() - cls.__starter_threads.remove(thr) - - for srv in cls.__servers: - srv.stop() - cls.__is_started = False - - @classmethod - def allStats(cls): - res = ["Servers %s" % ('stopped', 'started')[cls.__is_started]] - res.extend(srv.stats() for srv in cls.__servers) - return '\n'.join(res) - - def _close_socket(self): - close_socket(self.socket) - -def replace_request_password(args): - # password is always 3rd argument in a request, we replace it in RPC logs - # so it's easier to forward logs for diagnostics/debugging purposes... - if len(args) > 2: - args = list(args) - args[2] = '*' - return tuple(args) - -def dispatch_rpc_log(logger, prefix, msg, depth=None): - indent='' - indent_after=' '*len(prefix) - for line in (prefix+pformat(msg, depth=depth)).split('\n'): - logger.debug(indent+line) - indent=indent_after - -def dispatch_rpc(service_name, method, params): - """ Handle a RPC call. - - This is pure Python code, the actual marshalling (from/to XML-RPC or - NET-RPC) is done in a upper layer. - """ - try: - rpc_short = logging.getLogger(__name__ + '.rpc_short') - rpc_full = logging.getLogger(__name__ + '.rpc_full') - rpc_short_flag = rpc_short.isEnabledFor(logging.DEBUG) - rpc_full_flag = rpc_full.isEnabledFor(logging.DEBUG) - if rpc_short_flag or rpc_full_flag: - start_time = time.time() - if rpc_full_flag: - dispatch_rpc_log(rpc_full,'%s.%s:request '%(service_name,method), replace_request_password(params)) - - result = ExportService.getService(service_name).dispatch(method, params) - - if rpc_short_flag or rpc_full_flag: - end_time = time.time() - if rpc_full_flag: - dispatch_rpc_log(rpc_full,'%s.%s:reply time:%.3fs '%(service_name,method,end_time - start_time), result) - else: - dispatch_rpc_log(rpc_short,'%s.%s time:%.3fs '%(service_name,method,end_time - start_time), replace_request_password(params), depth=1) + logconfig = tools.config['log_handler'] - return result - except openerp.exceptions.AccessError: - raise - except openerp.exceptions.AccessDenied: - raise - except openerp.exceptions.Warning: - raise - except openerp.exceptions.DeferredException, e: - _logger.error(tools.exception_to_unicode(e)) - post_mortem(e.traceback) - raise - except Exception, e: - _logger.error(tools.exception_to_unicode(e)) - post_mortem(sys.exc_info()) - raise + logging_configurations = DEFAULT_LOG_CONFIGURATION + pseudo_config + logconfig + for logconfig_item in logging_configurations: + loggername, level = logconfig_item.split(':') + level = getattr(logging, level, logging.INFO) + logger = logging.getLogger(loggername) + logger.setLevel(level) -def post_mortem(info): - if tools.config['debug_mode'] and isinstance(info[2], types.TracebackType): - import pdb - pdb.post_mortem(info[2]) + for logconfig_item in logging_configurations: + _logger.debug('logger level set: "%s"', logconfig_item) + +DEFAULT_LOG_CONFIGURATION = [ + 'openerp.workflow.workitem:WARNING', + 'openerp.http.rpc.request:INFO', + 'openerp.http.rpc.response:INFO', + 'openerp.addons.web.http:INFO', + 'openerp.sql_db:INFO', + ':INFO', +] +PSEUDOCONFIG_MAPPER = { + 'debug_rpc_answer': ['openerp:DEBUG','openerp.http.rpc.request:DEBUG', 'openerp.http.rpc.response:DEBUG'], + 'debug_rpc': ['openerp:DEBUG','openerp.http.rpc.request:DEBUG'], + 'debug': ['openerp:DEBUG'], + 'debug_sql': ['openerp.sql_db:DEBUG'], + 'info': [], + 'warn': ['openerp:WARNING', 'werkzeug:WARNING'], + 'error': ['openerp:ERROR', 'werkzeug:ERROR'], + 'critical': ['openerp:CRITICAL', 'werkzeug:CRITICAL'], +} # vim:expandtab:smartindent:tabstop=4:softtabstop=4:shiftwidth=4: