-
Notifications
You must be signed in to change notification settings - Fork 123
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Feature]: Lighter weight Java runtime like OpenJ9 #373
Comments
I can't say I know anything about OpenJ9/IBM Semeru so I can't really be sure about comparisons to OpenJDK JRE 17, which we are using today. My main concerns would be things like:
I can put trying to put together a proof of concept together but it'd be something that would be on the back burner for me in all honesty. I know what I am getting with OpenJDK JRE and I know the support lifecycle because it's packaged with Ubuntu. |
I had a go at this on https://github.com/ktims/docker-omada-controller/tree/openj9-testing, initial results are promising, I see more than 45% reduction in container memory utilization. It seems to work fine though I haven't tested it extensively or with a 'real' workload. OpenJDK:
OpenJ9:
What's more, OpenJ9 is aware of container memory restrictions, so if I'm really rude to the container and only give it 512m RAM, it can be even more aggressive:
|
Interesting. Do you have example install code so I could take a look as well? Significant memory reduction could be really interesting, especially for the lower powered systems a lot of people tend to run this on. Also, I'd like to do some tests with MongoDB and the controller running as separate processes to get metrics from the only the controller to remove that variable. |
Not sure what you mean about install code, everything you need to try it should be in my
That's all that was required to get it up and running. OpenJ9 definitely feels subjectively slower when it's cold, but once warmed up it actually seems to outperform OpenJDK based on pageload timing, which is pretty surprising to me. Test based on the
|
This looks very promising for us 'home' users. I have an RPI4 2GB which I use to host a few other containers and sits about 1GB usage, but the current full-OpenJDK image just pushes it too far and hits OOM issues. Limiting the container memory to 800MB stops the OOMs but the controller software becomes unusable. So I'm running on a desktop host instead, which is not ideal because I would prefer to use a low power device so the controller can be left running 24/7. If @ktims testing is anything to go by I would be able to run the OpenJ9 image and still have a bit of memory spare. Apologies I cannot assist with development, but I am happy to assist with testing it in a 'production' environment if it gets to that stage (1 switch and 3 EAPs with approx 30 clients max) |
Sorry for the lack of progress so far. I've added this to my backlog of things to look at further. |
Nice.. Good to hear.. |
I started some work on a custom base image for OpenJ9 because I prefer to have consistency & control over the ability to patch the base image. It's nothing crazy (Dockerfile for this), just taking the OpenJ9 images, grabbing the JRE + the share classes cache and putting it in an image (on Docker Hub). OpenJ9 only has arm64 and arm64 builds available but I don't really see that as a problem as the armv7l images are already doing something different as it is today. My builds really aren't doing anything different than the I hope to get a chance to test this out later today with an actual build so I can do some comparisons myself. If I do, I'll update here and probably put up a branch of what I have. |
OK, so I have a branch https://github.com/mbentley/docker-omada-controller/tree/openj9 that seems to work (comparison from master). I just built a couple of test images for amd64 and arm64: |
Thanks @mbentley if we run a container from this image can it retain the same volumes as a container that was running the OpenJDK image? Or would it be better to set this up as a "fresh" container and use the controller migration within the omada controller software to move devices across instead? |
It should be fine but I would make sure to take backups before hand (you should be taking regular backups anyway - autobackups are built into the controller software unless you haven't enabled them). Keep in mind, this isn't merged into master as I need to do some more testing so there may be some changes but I intend of them to not be breaking changes. |
Yeah for sure, always backup haha. Do you envisage having two images once this is work is complete? A "full" image plus a "lite" image, so to speak? Or are you actually thinking you'll permanently switch to using OpenJ9 going forward? |
I did some comparisons between the normal OpenJDK, OpenJ9, and OpenJ9 with
For JVM startup times on a clean install:
So startup times were pretty close to identical on my server. |
Ideally, I would like to not add additional images to build as I am currently building 26 images just for the Omada Controller versions that are "supported". Add in those that are "archived" (I don't build these daily), it's 65. This includes amd64, arm64, armv7l, and the versions with chrome for report generation. My thought being that OpenJ9 is basically an extended OpenJDK so I would hope that there are no regressions or complications from switching. |
Did you have any 'load' on the controllers during these tests? ie: Were there actually any devices adopted into the controllers, any traffic flowing on the devices under their control, etc? |
No, this was just for overall startup of a brand new controller, no devices under management. Wasn't yet ready to try anything with my own instance yet. |
You may want to try populating your own shared classes cache in your image build step rather than copying the one from the original containers (assuming I understood what you wrote earlier about creating your own containers). If you do an application startup inside the build step, it should create a custom shared classes cache for your app inside your container that can then start your container more quickly. Even better if there is a way to run some load in that build step, because then you'll be able to get JIT compiled code cached right into your container image (if you keep that Great to see people getting value from Eclipse OpenJ9 !! Best of luck! |
Thanks @mstoodle for the tip! That's correct, I am just pulling the shared classes cache from the Docker Hub image so it sounds like I have some playing around to do to see what might be possible to optimize the cache. I have limited information about the software app itself. If you don't mind me asking, one thing I would be curious about would be if one approach could be to use a shared classes cache that is persistent in a directory outside of the container that is read/write. The first startup of the app wouldn't be optimized but I would image that subsequent startup and running of that app may? Is that an approach worth investigating or would that be an anti-pattern? Just curious as getting this app started as a part of a build step might introduce some complexity on my end that might be a bit funky and resource intensive considering this app needs and auto-starts MongoDB as well. I'll also see if I can locate the documentation on shared classes caching works as I will admit I haven't even looked yet. |
Hi @mbentley . You can configure the cache to reside on a docker volume if you like, but it gets troublesome to manage (at least in general deployments; not sure if the complexity would be there in your case). But there are advantages to having the cache inside the container. You can make the prepopulated cache layer read-only which speeds up the access to it. If you have people who build image layers on top of yours, they can add their own layer to the shared cache too (it's designed to work alongside how docker layers work). Quick references to the shared cache doc: https://eclipse.dev/openj9/docs/shrc/ or, if you prefer blogs you can look at some here: https://blog.openj9.org/tag/sharedclasses/ . If you have questions, you can @ mention me here and I'll try to respond. |
Do you have a recent version of this test image? |
Sorry, I haven't had to chance to follow up on anything further but I was able to build some new images using the latest version just now:
I've done no further testing on them yet but I assume they start up :) |
Stable also for me. Already using it for a month.. |
At this point, I would like to work on better understanding a shared classes cache pattern that makes the most sense for how the app runs in a container. I see a lot of experimentation in the future to make that happen. |
I am pretty new to docker, to run the openj9 container I would just replace the container name from the default in the build command with the new one and keep all the port allocates and everything the same? |
Correct. And to be clear, I am manually building this image right now so it's not getting automatic updates at the moment. I don't expect there to be issues but just FYI. Make sure you're taking regular backups of your persistent data. |
I am actually running it on a Raspberry Pi 4 and running the omada SDN on
bare raspbian, so going to save that microsd card as the backup. Just
wanted to give docker another go (was having issues with your docker image
and performance) which is why I went to a bare install. Looks like the
memory issue was what was causing it.
…On Wed, May 1, 2024 at 9:51 AM Matt Bentley ***@***.***> wrote:
I am pretty new to docker, to run the openj9 container I would just
replace the container name from the default in the build command with the
new one and keep all the port allocates and everything the same?
Correct. And to be clear, I am manually building this image right now so
it's not getting automatic updates at the moment. I don't expect there to
be issues but just FYI. Make sure you're taking regular backups of your
persistent data.
—
Reply to this email directly, view it on GitHub
<#373 (comment)>,
or unsubscribe
<https://github.com/notifications/unsubscribe-auth/AE4D4LPBIRGRK63DCFETHULZADXO5AVCNFSM6AAAAAA7R5PD4KVHI2DSMVQWIX3LMV43OSLTON2WKQ3PNVWWK3TUHMZDAOBYGQ4TKNBQGY>
.
You are receiving this because you commented.Message ID:
***@***.***>
|
I did not expect such an improvement, but here we are, it seems to be running at half of memory than previously (1GB vs 2GB) Good idea @paulschmeida and good job @mbentley, Would love to see this also for 5.14 :) |
So I just made a few changes which should make this able to be merged so that the Right now, A quick build command that will work for now is: docker build --pull --progress plain --build-arg BASE="mbentley/openj9:17" --build-arg INSTALL_VER=5.14.26.1 --platform linux/amd64 -t mbentley/omada-controller:test-openj9-amd64 -f Dockerfile.v5.x . |
This is awesome... I recently switched back to docker (to many issues
running the controller native on debian) but have missed openj9 while
running the latest beta
…On Tue, Oct 1, 2024, 09:13 Matt Bentley ***@***.***> wrote:
So I just made a few changes which should make this able to be merged so
that the install.sh script will use OpenJ9 if found (need to specify the
correct base image that already has OpenJ9 as seen below in the example
build command) or fall back to OpenJDK 17. Once I get this merged, I can
start to do some auto builds of an OpenJ9 tag but I don't yet want to make
it the default version as it seems like there are still optimizations that
can be done with the shared classes cache.
Right now, JAVA_TOOL_OPTIONS is being set in the base image layer
<https://github.com/mbentley/docker-base-ubuntu/blob/master/Dockerfile.openj9#L11>
and at this point, I am pretty sure I just mostly coped that from here
<https://github.com/ibmruntimes/semeru-containers/blob/1212e4fe213cb5b4c65eb260ccbbc40a7eadfb5d/17/jre/ubuntu/focal/Dockerfile.open.releases.full#L61>
.
A quick build command that will work for now is:
docker build --pull --progress plain --build-arg BASE="mbentley/openj9:17" --build-arg INSTALL_VER=5.14.26.1 --platform linux/amd64 -t mbentley/omada-controller:test-openj9-amd64 -f Dockerfile.v5.x .
—
Reply to this email directly, view it on GitHub
<#373 (comment)>,
or unsubscribe
<https://github.com/notifications/unsubscribe-auth/AE4D4LLZVTU6ZZT6KQ77KJLZZKNYJAVCNFSM6AAAAAA7R5PD4KVHI2DSMVQWIX3LMV43OSLTON2WKQ3PNVWWK3TUHMZDGOBVG43TOMJRGM>
.
You are receiving this because you commented.Message ID:
***@***.***>
|
So now that I just merged #479, I'm trying to determine what which images are worth building with OpenJ9 to be test images. Right now, I am thinking that the latest 5.14 and 5.14 beta might be worth doing but I know that the current 5.14 GA version has random issues documented in #418 with the app itself so maybe I could also include 5.13 on that. Thoughts? |
I agree with 5.13/5.14/5.14beta |
I'm waiting on automated build on OpenJ9 and will move to use it immediatly it happens. I like to be on the bleeding edge, but I do like to automate things with watchdog, so once we have any automated builds, I'm gonna move to use it. Now running 5.13 since the 5.14 bug preventing me from updating, but once they release a new 5.14 that fixes it, will move to that. |
So I agree that 5.14/5.14/beta releases are enough for OpenJ9 in my opinnion! :D |
nice.. |
OK, here are the autobuild tags that I just added. I still need to add something to the readme but this will work for now: 5.13multi-arch
arch specific
with chromium
5.14multi-arch
arch specific
with chromium
betamulti-arch
arch specific
with chromium
|
Just switched to openj9, and while it's still painfully slow to use, multisecond page loads etc, at least the memory footprint is lower, and yes, nothing went differently, i'm still using 5.13 so everything just worked. I have 3 devices, and about 20 clients. OpenJDK
Openj9
Overall, really good update! Can't wait for 5.14, and hopefully we can get some additional speed improvements with shared cache implementations. I would help if i would know how. But I'm total newbie with java stuff. I have this in my logs now, but
|
The lack of shared cache is why this issue is still open - right now, there is no SCC being generated. I have a WIP hack based on the method I see here https://github.com/ibmruntimes/semeru-containers/blob/1212e4fe213cb5b4c65eb260ccbbc40a7eadfb5d/17/jre/ubuntu/focal/Dockerfile.open.releases.full#L69-L110 but I have yet to commit that to the |
I don't know, I run it ( |
I'm not sure if this is related to this, but I'm running my Omada on Rasp5 with 8GB of RAM, and for some reason every single page load takes multiple seconds, loading the organization front page takes about ~10-20 seconds. The UI is really sluggish. I don't really mind, since I'm only using it when I need to actually accomplish something, but it is weird. Maybe it's an ARM issue? I can make a separate issue to track this if it would be helpful and provide a lot more info. |
And I always thought it was just Omada issue since they have acknowledged that even their HW Controllers are slow for some: https://community.tp-link.com/en/business/forum/topic/639500 |
I'm curious if you've done any benchmarking of your storage speed on your Raspberry Pi 5. I wouldn't think that the CPU or memory for a Pi would cause significant slowdowns and can only really think of storage speed maybe being the bottleneck outside of some potential bug in the software. |
Yeah, I've been running nVME SSD for ~5 months now with these speeds: Write 544 MB/s, Read 842 MB/s. I too thought that nvme upgrade would have helped with omada running sluggish, but a'las, that didn't manifest. |
Well there goes that idea 😆 |
This might be one of the most disgusting things I have done in a Dockerfile in a long time but I thought I would at least commit what I had so far on a RUN build step which is extremely hacky and probably not how I would actually do this. It seems like it populates the cache but I'll have to see as I only had time to see that it basically outputs a number and that's all. I never tested to see if it was actually using it a startup time. |
There is no |
There is an 5.14-openj9 tag. I need to update the main README as I am regularly building 5.13, 5.14, and the beta version with OpenJ9. |
… On Tue, Nov 19, 2024, 08:45 Matt Bentley ***@***.***> wrote:
There is an 5.14-openj9
<https://hub.docker.com/r/mbentley/omada-controller/tags?name=5.14-openj9>
tag. I need to update the main README as I am regularly building 5.13,
5.14, and the beta version with OpenJ9.
—
Reply to this email directly, view it on GitHub
<#373 (comment)>,
or unsubscribe
<https://github.com/notifications/unsubscribe-auth/AE4D4LIICBCGL2YDDIHCIX32BM6GXAVCNFSM6AAAAAA7R5PD4KVHI2DSMVQWIX3LMV43OSLTON2WKQ3PNVWWK3TUHMZDIOBVG43DINJTGU>
.
You are receiving this because you commented.Message ID:
***@***.***>
|
Hm, I thought I tried that, but apparently I didn't. Tried it now and it works*, thanks! * After some issues with MongoDB being killed because I had a memory limit set. It seems |
Is OpenJ9 still preferred if I'm running this on a powerful machine? I'd prefer better performance overall. |
I wouldn't really say there is any particular noticeable difference on a powerful machine but I haven't done extensive testing. At the moment, it seems to be more about memory usage when running where the OpenJ9 containers will use less. |
I will say I was using OpenJ9 for a while and switched back to just the default Beta (after using the OpenJ9 beta of 5.15.8.1) and the WebUI is much faster/snappy with the default Java build vs the OpenJ9. I think for the less RAM that OpenJ9 uses, the performance decrease isnt worth it. |
What problem are you looking to solve?
Since a lot of people run this software for their home networks with just a couple of APs, having a docker container with full-blown vesion of Java seems like an overkill.
Describe the solution that you have in mind
I propose the switch to a lighter wait version of JRE, one that's built specifically for containers. Now bear in mind I'm not a Java developer, but I've been told that OpenJ9 from IBM Semeru for example is more lightweight and can run the same java apps at 1/3 memory footprint and lower CPU usage.
Additional Context
No response
The text was updated successfully, but these errors were encountered: