Leaderboard


Popular Content

Showing content with the highest reputation since 09/21/2018 in all areas

  1. 3 points
    I would love to see LM implement a new feature for taking a built-in, self prescribed, action on an alert. To minimize any exposure that LM might have in an action gone awry, the actions taken could occur as the result of a script that one could upload into the Escalation Chain. Ideally you could define multiple actions or multiple retries on an action and whether that occurred before or after the recipient notification in the notification chain. This would allow for very basic alerts (disk, service restarts, etc) to be resolved programatically. Also being able to support various scripting languages such as PowerCLI, Ansible, etc would allow for some very creative ways to integrate with solutions such as VMWare or Ansible Tower for very complex actions to be crafted by more expert skill level folks.
  2. 3 points
    We use the Debug Console a lot, and usually to debug an issue on a particular device, which means more often than not, we access the Debug Console from the device in question's Raw Data view. We would like for there to be a token (##THIS## or ##DEVICE##) that could be used with Debug Console commands. So if we were to issue the command !ping ##THIS##, the console would automatically interpolate ##THIS## with the IP address of the device on which we opened the Debug Console. It's a small productivity gain, but when you're doing this a lot it makes a big difference.
  3. 2 points
    Most often, when people export to a csv or excel format their intent is to receive table data in a tabular format because they're going to pivot it out, chart it, or conduct some sort of analytics/BI function. It would be nice if your csv exports didn't require manipulation of the data to remove erroneous data/whitespace for consumability as a table datasource. This is specifically a problem in Website Overview reports.
  4. 2 points
    So I know this is an old thread, and the above community locator isn't really needed now that the EIGRP peers datasource is in the LM repo.... but I thought I'd post here in case someone else hit a similar issue and wanted a fix. The built-in groovy discovery script was using HEX addresses for the peer addresses in our instances, and it was a pain to decipher them every time one went down. So I added a hex-to-decimal conversion to the discovery script so that he instance names look like "172.20.0.50" instead of "AC:14:00:32" Here is the updated block of the discovery script // get all IP addresses of connected peers. peerAddr_walk.eachLine { line -> regex_pattern = "(${peerAddr_OID})" + /.(.*)\s=\s(.*)/; regex_match = ~regex_pattern line_match = regex_match.matcher(line) handle = line_match[0][2] val = line_match[0][3] //val here may be SNMP data in format of "Hex-STRING: AC 14 00 32" and arrive as AC:14:00:32 instead of an ip address in the instance name if (val.contains(":")) { tempaddr = val.split(":") def newaddr = [] tempaddr.each { hexint -> newaddr.add(Integer.parseInt(hexint, 16)) } val = newaddr.join(".") } Hope it helps someone else
  5. 2 points
    I think it would greate if you add some headers to your mails. This will helps to mail program to create conversation for every alert and clean message for it. Now we have only separate messages: LMD... critical - Host1 Ping PingLossPercent LMD... critical - Host2 Ping PingLossPercent LMD... ***CLEARED***critical - Host2 Ping PingLossPercent LMD... ***CLEARED***critical - Host1 Ping PingLossPercent In my opinion, it will better if this message will create conversation for every alert: LMD... ***CLEARED***critical - Host1 Ping PingLossPercent LMD... critical - Host1 Ping PingLossPercent LMD... ***CLEARED***critical - Host2 Ping PingLossPercent LMD... critical - Host2 Ping PingLossPercent As I know, the header is Thread-Index https://excelesquire.wordpress.com/2014/10/17/use-excel-to-count-the-number-of-emails-in-each-email-chain/ https://stackoverflow.com/questions/5506585/how-to-code-for-grouping-email-in-conversations
  6. 1 point
    Would LM consider adding more Checkpoint locations in the UK or Europe please? As a UK customer we get poor response times from Sydney and Singapore from the UK. US locations also are not ideal. Could we also have a feature to add our own LM Collectors as checkpoint locations too please?
  7. 1 point
    Hi Sarah, thanks for the swift response, apologies but let me explain our need more clearly... We need to perform website checks from the internet from multiple locations. While LM has 6 internet locations only 1 is close to the UK and the more remote sites appear unreliable. We can only set a single timeout/download time to cover all monitoring locations but even at 30,000ms we still find that we are getting drops and false alarms. Having a number of reliable checkpoint locations is crucial for us so we can protect against false positives and can obtain a reliable reading of how responsive a site actually is. So we were wondering if LM intend to open any further checkpoint locations in the near future - hopefully in the UK or Europe ideally? The [only] alternative we have currently is to build out our own "Collector virtual machines" in data centres around the UK and Europe and use the "Internal Web Check" functionality as you say.
  8. 1 point
    DataSource scripts are meant to execute quickly, so just be careful that you're "task" logic is not doing too much work. May be better if the task logic is in another script that the DataSource script launched, and then the DataSource script ends.
  9. 1 point
    Alert processing happens outside the detection point (in "the cloud") -- there need to be triggers to an event handler that operate in the collector context. One possibility would be to create datasources that don't actually collect data, but do the check and repair operation, with a datapoint as a side effect. It would be easier if datasource code could cross-reference other datasource/instance datapoints without having to replicate the same API code into each (e.g., code library support), but it is feasible. Triggers would be much cleaner.
  10. 1 point
    While looking through the configuration of alert templates, I see that the various types of alerts and that each can have a customized template defined. While that's great, it would be even better if a custom template could be defined on a per device / datasource/ eventsource/ etc. basis to allow for better communication with the folks getting the alerts.
  11. 1 point
    I'm currently new to LogicMonitor. I think LM has done a pretty good job on their monitoring tool. I love how we could manage our network devices thru SSH remote access from the management console. Without LM, we would have to VPN to our internal network, and then SSH to the network devices from there. One downside about LM is that we couldn't manage network devices thru HTTPS (GUI). Nowadays, new technologies like firewalls are managed by HTTPS/SSH. I would love to have HTTPS remote access right from the management console. It would be much easier and faster to gain access to the GUI interface. I have tested Auvik monitoring software on the HTTPS, it was nice and smooth. This would be one of the nice feature that LM should provide. Thanks, Pao
  12. 1 point
    The alert "template" system has no way to reference another instance datapoint value currently. Why would this be good? If you have a datapoint that alerts for status, you want to insert the value of the thing you care about, not the status value. I see the contortions LM datasource developers have gone through to workaround this (e.g., Cisco Nexus Temperature). Please make it possible to reference datapoints at least within the same datasource within alert templates. There are many other issues with alerts, but I will stop here for now :).
  13. 1 point
    Hi, Our team recently has certain error scenarios found in multiple production sites. As of today we're monitoring specific exception (via keyword match or Regex expression) via LogicMonitor and trigger alert to be generated. This solution has few drawbacks: 1. Requires us to know ahead what're the specific exception(s) to monitor in each log file (e.g. Tomcat, ActiveMQ) 2. Requires us to download all the logs from each production site that has this issue (some of our customers requires VPN/Secure access and it's very inefficient to download these logs from each site to analyze) Our team then run a quick log streaming POC and discovered datadog is one of the vendors that provides a decent log streaming solution (to the cloud) and allow us to search & perform analytics (see https://www.datadoghq.com/log-management/). It'll be great if LogicMonitor can implement something similar to enable us to elasticsearch these logs in the cloud to enable faster troubleshooting analysis. Thanks & Best Regards, Horace
  14. 1 point
    Collector groups were added recently, and are detailed here: https://communities.logicmonitor.com/topic/637-collectors Now let's expand upon that functionality...What if collectors be grouped dynamically? Identically to how Devices dynamic groups work, could I assign properties to collectors, then build dynamic groups based from the properties? Ways that envision sorting collectors: Production/test, primary/backup, collector number (1-99, 100-199, 200-299, etc.), zip code, time zone, alphabetical. In each of these cases, this would give full flexibility on how collector upgrades are scheduled. Currently if we have a mass collector upgrade to a new General Release, it can be a little painful to manage so many upgrades running simultaneously (or at least in very short order). I am most interested in being able to split them up into primary, backup and single collector groups. This way, I know that it's pretty safe to upgrade the collectors that have backups after hours, since there is another collector to failover to. And I surely want to be available staff-wise if I am doing upgrades for those collectors that have no backup collector. Close behind sorting into primary/backup/single is the need to sort them by customer (which currently works fine). The issue is that you can't put a collector into more than one group, which precludes from even setting up these to items manually.
  15. 1 point
    Thanks Kerry! We're digging into the new sources now and they look very useful. I know we just got these but I'll go ahead & submit my wish list for additional enhancements: Support for servers running more than one SID (as is, I have to modify all these to support a property listing multiple SIDs) Oracle ASM support Oracle RAC metrics
  16. 1 point
    Hi, Per discussion with Russ G. & Kenyon W. & Jake C. yesterday, I would like to submit this as a feature request to the DEV team and see whether there is any way to add this feature into future roadmap. In short, it'll be great if end user can configure multiple incident/alerts into 1 group and generate only 1 alert (with highest severity). Here is an example of Tomcat being shutdown which shows a number of alerts generated: 1. Tomcat shutdown ‘critical’ alert is generated (1 alert) 2. ActiveMQ consumer count of specific queue alert has reached zero ‘Error’ alert (about 10-12 alerts for our case) In this case end user would like to be able to configure such that LM will consolidate all alerts into one critical alert (i.e. all AMQ 'Error' alerts are cleared)? I saw something like this in PagerDuty and must say it’s a great feature to have in LogicMonitor to reduce # of alerts being processed by the TechOps team: https://www.pagerduty.com/blog/alert-triage/ Thanks & Best Regards, Horace
  17. 1 point
    I know that it is not exactly recommended/reliable to use a 1GB/1CPU Core machine to monitor...but it seems that installing a "nano" sized collector on a t2.micro AWS instance and having it just monitor itself brings the AWS instance to a screeching halt. I am seeing that when the collector is running, top shows that CPU pegs to 100% almost nonstop. Memory is not hit quite as bad..but it does get up there to use 500mb+ But the CPU load average is 5+ cores and it makes the system unusable. Sometimes this causes the instance to throw status alerts & even crash. Question: Has anybody been able to tweak the wrapper.conf etc files to make the collector CPU load less demanding?
  18. 1 point
    Do you mean within the web GUI? This functionality exists in the API: https://www.logicmonitor.com/support/rest-api-developers-guide/v1/sdts/get-sdts/
  19. 1 point
    A workaround using the !groovy command: def sout = new StringBuilder(); def serr = new StringBuilder(); // Shell command for your OS def proc = 'cmd.exe /c tracert www.google.com'.execute(); proc.consumeProcessOutput(sout, serr); // Adjust timeout as needed proc.waitForOrKill(20 * 1000); println "out> $sout err> $serr"
  20. 1 point
    We have a use case to show "Response Times" from a subset of configured Websites. Ideally I'd like this to be in the Big Number widget. We also want to able to chart a subset of my Websites' response times over time in the Chart widget. Anyone found a useful workaround to achieve this? Would LM consider "upgrading" widgets to allow the presentation of Website data? Currently only the SLA widget seems capable of handling Website data.
  21. 1 point
    HCPFGA The default LogicMonitor datasource names the instances in a strange way and then alerts for events that have already completed. I've added a better instance naming convention that clearly identifies the event that will occur and when. I also put in logic to detect if the scheduled event has already taken place to prevent unnecessary alerting.
  22. 1 point
    Hey LM team, There are a number of options to customize what is in a service alert as well as a data source alert, but config source? Nope, can't modify what gets sent. All it says is that a change was made and gives you a link to the device. A couple things I think would be beneficial as a new feature, 1. Being able to put what lines were added or removed right in the alert. For simple configs you can then tell right away if you need to take action or not. 2. Have the link in the alert actually bring you to the diff view of the current and last config state. While having to click through the multiple links and then slide the diff view slider is fairly painless, at 3am on my phone it can add to the sleepy frustration level. Basically just bringing some of the flexibility that alert templates have for data sources and services to the config sources. Thanks!
  23. 1 point
    I'm not having any luck with Glob filters. Also, just a suggestion. I get that you have a mixture of Linux / Windows admins. Me being a windows admin, I'm going to be a bit bias in saying that I find regex style patterns to be overly convoluted, and honestly wish you guys simplified your operating / pattern matching language. Certainly nothing wrong with offering regex as an option, but its not my go to langue for pattern matching, or its rather a language of last resort that I use. For example: I have a group structure that looks like this Windows Servers\ Exchange SQL IIS DHCP AD Cisco\ ESA WSA Switch FW I want to show all devices in the Windows server, except IIS and I want to show only ESA in Cisco. Doing something akin to.... Select * from groups where (name like *windows* and not like *IIS*) or (name like *ESA*) That's simple SQL language, and IMO far easier to read / understand. Even if it's not SQL based language, I think working on more friendly (readable) language would be a huge benefit. Even something more simplistic like (like *windows* and notlike *IIS*) or (like *ESA*). Maybe even take it a step further and offer multiple query languages that we can pick from. Sometime regex IS the only language that can get it done. And while I realize regex more often than not can do what SQL cannot, 99% of the examples I'm showing, don't require the complexity of regex.