Or any other log files/output? I’m open to any solution, but what I would like…

…is something where I can just click on a word or select some text and say “filter that out”

Something that colors different log levels differently, preferably automatically.

Something that can parse the “columns” and give me a nice quick list of values, like different unit names to filter out/solely include.

Something that lets me choose a time and go there. Something that lets me select only a specific timeframe of logs.

I know this can probably be done by going in/out of journalctl, recalling the last command and adding specific filter options… but it just feels slow. It’s so many keypresses when I could just right click on the word and -> “Filter out/Search for” or something.

  • mosiacmango@lemm.ee
    link
    fedilink
    arrow-up
    2
    ·
    edit-2
    1 hour ago

    Sounds like you want a siem like Wazuh. Its agent can collect journald logs from any number of systems. It also has a gui you can interact with to parse logs.

  • stembolts@programming.dev
    link
    fedilink
    arrow-up
    4
    ·
    edit-2
    3 hours ago

    tbh my go to command is just… journalctl -fe -u service

    ex :
    journalctl -fe -u jellyfin
    journalctl -fe -u nordvpnd

    so I’d also like to know the answer to this question. my other go to is dumping journalctl to text files and parsing with grep and awk and creating my own reports with that parsed information.

  • aleats@lemmy.blahaj.zone
    link
    fedilink
    English
    arrow-up
    4
    ·
    3 hours ago

    I don’t know of any graphical tools that let you do this, but generally, if you want to search for specific terms/times/commands or anything of that sort, piping journalctl into grep (and optionally grep into less) is pretty effective at finding stuff.

  • blobjim [he/him]@hexbear.net
    link
    fedilink
    English
    arrow-up
    2
    ·
    3 hours ago

    I wish there was something nice like that too.

    In the server world that would usually involve doing something like sending the journal data to Elasticsearch using an Elasticsearch integration. But that involves setting up an Elasticsearch server and Kibana and so on which is very unwieldy for a desktop computer. It does work pretty well though in terms of filtering. But it also stores the data internally in indexes to speed up search.

    Of course journald has a seemingly simple C API but writing code is a lot of work. There are probably API bindings for various languages.