17
Raptor Lake efficiency cores in Linix with Docker?
(lemmy.nowsci.com)
From Wikipedia, the free encyclopedia
Linux is a family of open source Unix-like operating systems based on the Linux kernel, an operating system kernel first released on September 17, 1991 by Linus Torvalds. Linux is typically packaged in a Linux distribution (or distro for short).
Distributions include the Linux kernel and supporting system software and libraries, many of which are provided by the GNU Project. Many Linux distributions use the word "Linux" in their name, but the Free Software Foundation uses the name GNU/Linux to emphasize the importance of GNU software, causing some controversy.
Community icon by Alpár-Etele Méder, licensed under CC BY 3.0
A Docker container is a security framework. The process running "inside" the container is just a Linux process like any other.
So, as I understand it, the performance will be identical to a process that is running "outside" a container, subject to the overhead associated with any security restrictions.
If you want to protect the system from untrusted software with containers be careful. Containers and images are mostly an abstraction tool to run and control the applications. Not saying it's not possible, it's just easy to make it insecure.
Yes, this is the case, I'm more wondering about kernel support for CPU assignment as it relates to those processes.
I think that what you're looking for is "CPU affinity", but that is not something I know anything about.
In the 40+ years I've been playing with computers, I've always let the OS worry about where and when to run a process and only rarely do I
renice
a process that needs to run, but not at the expense of everything else.Agreed, just want to make sure the kernel can handle resourcing for two different types of cores. I know there was a time (recent) where it couldn't. Others have said 6.x is the key.