Zolfa 241bf1219d | 11 months ago | |
---|---|---|
.github/workflows | 2 years ago | |
logzio | 1 year ago | |
tests | 2 years ago | |
.gitignore | 7 years ago | |
.travis.yml | 3 years ago | |
.whitesource | 6 years ago | |
LICENSE | 9 years ago | |
README.md | 1 year ago | |
__init__.py | 8 years ago | |
build.sh | 7 years ago | |
requirements.txt | 1 year ago | |
setup.cfg | 7 years ago | |
setup.py | 11 months ago | |
tox.ini | 3 years ago |
This project has been forked from The Logz.io Python Handler, version 4.0.1. After using the logz.io hosted and managed services to collect and visualize the logs of our Python appliances, we moved to a self-hosted on-premise OpenSearch cluster.
To reduce the burden of adapting our applications, we modified logzio-python-handler
to send the log stream directly to our OpenSearch
cluster. This fork work as a drop-in replacement for logzio-python-handler
, the name of classes is unaltered at the moment.
Instead of the parameters url
and token
the LogzioHandler constructors now requires host
, port
, username
, password
of the
OpenSearch cluster.
Only basic HTTP auth is supported at the moment, and a behavios has been hard-coded to produce new indices for every month.
Deprecation announcementVersion 3.0.0 of this project ends support for Python 2.7, 3.3, and 3.4. We recommend migrating your projects to Python 3.5 or newer as soon as possible. We'll be happy to answer any questions you have in a GitHub issue. Thanks! |
---|
This is a Python handler that sends logs in bulk over HTTPS to Logz.io. The handler uses a subclass named LogzioSender (which can be used without this handler as well, to ship raw data). The LogzioSender class opens a new Thread, that consumes from the logs queue. Each iteration (its frequency of which can be configured by the logs_drain_timeout parameter), will try to consume the queue in its entirety. Logs will get divided into separate bulks, based on their size. LogzioSender will check if the main thread is alive. In case the main thread quits, it will try to consume the queue one last time, and then exit. So your program can hang for a few seconds, until the logs are drained. In case the logs failed to be sent to Logz.io after a couple of tries, they will be written to the local file system. You can later upload them to Logz.io using curl.
pip install logzio-python-handler
Travis CI will build this handler and test against:
We can't ensure compatibility to any other version, as we can't test it automatically.
To run tests:
$ pip install tox
$ tox
...
[handlers]
keys=LogzioHandler
[handler_LogzioHandler]
class=logzio.handler.LogzioHandler
formatter=logzioFormat
args=('token', 'my_type')
[formatters]
keys=logzioFormat
[loggers]
keys=root
[logger_root]
handlers=LogzioHandler
level=INFO
[formatter_logzioFormat]
format={"additional_field": "value"}
args=() arguments, by order
Please note, that you have to configure those parameters by this exact order. i.e. you cannot set Debug to true, without configuring all of the previous parameters as well.
LOGGING = {
'version': 1,
'disable_existing_loggers': False,
'formatters': {
'logzioFormat': {
'format': '{"additional_field": "value"}',
'validate': False
}
},
'handlers': {
'logzio': {
'class': 'logzio.handler.LogzioHandler',
'level': 'INFO',
'formatter': 'logzioFormat',
'token': '<<LOGZIO-TOKEN>>',
'logzio_type': 'python-handler',
'logs_drain_timeout': 5,
'url': 'https://<<LOGZIO-URL>>:8071',
'retries_no': 4,
'retry_timeout': 2,
}
},
'loggers': {
'': {
'level': 'DEBUG',
'handlers': ['logzio'],
'propagate': True
}
}
}
Replace:
If you're using a serverless function, you'll need to import and add the LogzioFlusher annotation before your sender function. To do this, in the code sample below, uncomment the import
statement and the @LogzioFlusher(logger)
annotation line.
Note: For the LogzioFlusher to work properly, you'll need to make sure that the Logz.io. handler is added to the root logger. See the configuration above for an example.
import logging
import logging.config
# If you're using a serverless function, uncomment.
# from logzio.flusher import LogzioFlusher
# Say I have saved my dictionary configuration in a variable named 'LOGGING' - see 'Dict Config' sample section
logging.config.dictConfig(LOGGING)
logger = logging.getLogger('superAwesomeLogzioLogger')
# If you're using a serverless function, uncomment.
# @LogzioFlusher(logger)
def my_func():
logger.info('Test log')
logger.warn('Warning')
try:
1/0
except:
logger.exception("Supporting exceptions too!")
In case you need to dynamic metadata to your logger, other then the constant metadata from the formatter, you can use the "extra" parameter. All key values in the dictionary passed in "extra" will be presented in Logz.io as new fields in the log you are sending. Please note, that you cannot override default fields by the python logger (i.e. lineno, thread, etc..) For example:
logger.info('Warning', extra={'extra_key':'extra_value'})
LOGGING = {
'version': 1,
'disable_existing_loggers': False,
'formatters': {
'verbose': {
'format': '%(levelname)s %(asctime)s %(module)s %(process)d %(thread)d %(message)s'
},
'logzioFormat': {
'format': '{"additional_field": "value"}'
}
},
'handlers': {
'console': {
'class': 'logging.StreamHandler',
'level': 'DEBUG',
'formatter': 'verbose'
},
'logzio': {
'class': 'logzio.handler.LogzioHandler',
'level': 'INFO',
'formatter': 'logzioFormat',
'token': 'token',
'logzio_type': "django",
'logs_drain_timeout': 5,
'url': 'https://listener.logz.io:8071',
'debug': True,
'network_timeout': 10,
},
},
'loggers': {
'django': {
'handlers': ['console', ],
'level': os.getenv('DJANGO_LOG_LEVEL', 'INFO')
},
'appname': {
'handlers': ['console', 'logzio'],
'level': 'INFO'
}
}
}
Change
If you're sending traces with OpenTelemetry instrumentation (auto or manual), you can correlate your logs with the trace context.
That way, your logs will have traces data in it, such as service name, span id and trace id.
To enable this feature, set the add_context
param in your handler configuration to True
, like in this example:
LOGGING = {
'version': 1,
'disable_existing_loggers': False,
'formatters': {
'logzioFormat': {
'format': '{"additional_field": "value"}',
'validate': False
}
},
'handlers': {
'logzio': {
'class': 'logzio.handler.LogzioHandler',
'level': 'INFO',
'formatter': 'logzioFormat',
'token': '<<LOGZIO-TOKEN>>',
'logzio_type': 'python-handler',
'logs_drain_timeout': 5,
'url': 'https://<<LOGZIO-URL>>:8071',
'retries_no': 4,
'retry_timeout': 2,
'add_context': True
}
},
'loggers': {
'': {
'level': 'DEBUG',
'handlers': ['logzio'],
'propagate': True
}
}
}
Please note that if you are using python 3.8
, it is preferred to use the logging.config.dictConfig
method, as mentioned in python's documentation.
4.0.1
protobuf>=3.20.2
.setuptools>=65.5.1
4.0.0
3.1.1
3.1.0
3.0.0
python2.7
& python3.4
_flush_queue()
method (@hilsenrat)2.0.15
python3.7
and python3.8
2.0.13
pypy
and pypy3
(@rudaporto-olx)2.0.12 - Support disable logs local backup
2.0.11 - Completely isolate exception from the message
2.0.10 - Not ignoring formatting on exceptions
2.0.9 - Support extra fields on exceptions too (Thanks @asafc64!)
2.0.8 - Various PEP8, testings and logging changes (Thanks @nir0s!)
2.0.7 - Make sure sending thread is alive after fork (Thanks @jo-tham!)
2.0.6 - Add "flush()" method to manually drain the queue (Thanks @orenmazor!)
2.0.5 - Support for extra fields
2.0.4 - Publish package as source along wheel, and supprt python3 packagin (Thanks @cchristous!)
2.0.3 - Fix bug that consumed more logs while draining than Logz.io's bulk limit
2.0.2 - Support for formatted messages (Thanks @johnraz!)
2.0.1 - Added all to init.py, so support * imports
2.0.0 - Production, stable release.
1.X - Beta versions