Anyone using cloud based monitoring/logging services with Python logging module?
michael.vilain at gmail.com
Thu Jul 26 13:35:26 EDT 2018
I used DataDog. At the time, I didn't have any experience with python. I had to depend the company to make patches to things that didn't work. And anything custom, I just did a cron job instead of trying to integrate with their tool.
They marketed themselves as a monitoring company but they really collect a boatload of performance info that you can wade though to find out what's going on on the system being monitored. The alert setup and other monitor type things have to be done through the web GUI. I don't think you can code it though an API. You put whatever instrumentation you want on the server to be monitored, then clone those files and any local changes you made to their code.
It doesn't play well for setting up as an automated solution like stand alone monitoring packages. It's $15/node, which, if you 5000 nodes, you're talking real money.
Go with an on-prem solution rather than something that's cloud based. Those can be configured and deployed painlessly.
> On 26-Jul-2018, at 10:05 AM 🌞, Malcolm Greene <python at bdurham.com> wrote:
> Looking for feedback on anyone who's using a cloud based
> monitoring/logging service with Python's standard lib logging module,
> eg. services such as Datadog, Loggly, Papertrailapp, Scalyr, LogDNA,
> Logz.io, Logentries, Loggr, Logstats?
> Would appreciate hearing your experience, pros, cons,
> recommendations, etc.
More information about the Python-list