YouTubeWebsitePixelfed

Let me know if you want to mod any communities I’ve setup here on lemmy.world, thank you.


  • 0 Posts
  • 8 Comments
Joined 1Y ago
cake
Cake day: Jun 14, 2023

help-circle
rss

You don’t understand my point… I’m not going to trust someone that doesn’t know about what it was running at, which is why I asked for specifics on how much RAM it was using. I’m not trusting some random that says, “yeah bro, totally cool, not a problem even though I have no idea about what the RAM was at or even what I would consider to be high, but trust me bro, it’s okay. I am soooo vigilante on RAM usage on my server despite having no clue about the RAM usage, but trust me bro, I would know if it was leaking.”

Even after a restart of my NAS where every container was freshly started the RAM consumption wasn’t higher than a month later after the restart (without restarting any of the docker containers). And that is why I can with full conviction tell that my SearXNG instance didn’t leak memory.

No, you can’t, not unless you specifically paid attention to it, which you didn’t, and that is clear by you not knowing how much it was using. It’s possible that every time you checked your overall usage was just after a crash and the container restarting.

That’s the reason I asked you to give me specifics on the RAM… because without you being able to provide it, ruled out whether I trusted your opinion.


It doesn’t help, because I can’t trust it’s accurate, and neither should you. As I said, you can’t recall what the container was using, you can’t tell me any specific numbers like what you might consider to be too high, so how do you expect me to trust that you definitely, for sure would have noticed something that might have been as small as a 300MB swing?

Do you understand how this is not helpful in any way now?


No, I didn’t check because it never got so big to the point where it would become suspicious to me that something might be wrong.

And that’s all I needed to know, that is exactly what I thought, you didn’t take the time to verify anything.


It wasn’t much though and it never bloated, even when running for over a whole month.

Did you actually check on it, or did you just not notice a problem so you’re stating this? No offense, but you haven’t verified anything, and seem to just be recalling there wasn’t a problem.

I mean, what is “not much,” to you? Because I asked someone else what it was using who also thought it was running well for them, and after only 1 day of uptime it was idle at 350MB of RAM… which is way too high for an idle search engine in my opinion. Another thing, if you were running it in a docker container and had it set to restart=unless-stopped it could have been restarting without you even knowing about it.



Thanks for reporting back on it.

Yeah, even that, I just don’t understand why a search engine, sitting idle, consumes that much memory. The largest consumers of memory for me are: Omada Controller, Airsonic-Advanced, HomeAssistant, Lidarr, Sonarr, Paperless… all things that process a lot of data, or are written in Java, so it’s to be expected they use more resources.

Then I would see it grow to 600MB+ and occasionally crash, so I just decided to use public instances.


It was probably about a year ago now, so I probably should spin it up again. I think I’ll do that today.

How much RAM does yours use?


I went away from it because it seemed to have a memory leak and the docker container would eventually crash. I never truly investigated what was causing it.

I’ve been meaning to give whoogle a try:

https://github.com/benbusby/whoogle-search