The constant is defined between pairs of languages, and it tells you "how many bits does it take to emulate language A in language B". So it doesn't make sense to talk about "the" constant of a language, it's relative to what other language you are comparing it to.
So the answer is "it depends on the languages involved".
I thought it was talking about the overhead to describe an object in an absolute sense, but it turns out the constant is related to the difficulty of language emulation.
Practical programming languages are generally designed to try to reduce the Kolmogorov complexity of most common tasks, when pretty-much ignoring the additive constant for the language itself. This strongly encourages the additive constant for the language itself to be large, by adding a great many libraries useful for many types of tasks. For estimating the actual Kolmogorov complexity of a specific task for something like the Universal Prior, you're better off starting with one of the simplest Turing tarpits with low initial additive constant, and digging yourself out of the tarpit to the minimal extent actually required for that one specific task
I'm not sure the question makes sense.
For an analogy, consider the family of functions sin(x+c) for different values of c. Any of them is within an additive constant from any other (just take the constant equal to 2). But none of them has the "lowest additive constant", since they're all just shifted versions of each other.
So the constant value c is 2, then, at least once we discard the random variable x. I just checked with a calculator, and sin(2+2) is roughly equal to -0.76, rounded to 2 decimal places, and sin(4) got the same value, so the lowest constant is 2, unless it's lower than that, and I suspect that it still holds for all values of the random variable x.
Ah true, in this case we can take the maximum constant to be 2.
Let's try a different example. Consider the functions |x-c| for different values of c. Then if you take any two such functions, |x-a| and |x-b|, their difference is bounded by a constant |a-b| for all x. But it's not clear how to say which function has the "lowest additive constant".
Yeah, I'm really hoping Kolmogorov Complexity isn't like this, though when you say it's not clear that there's a minimum/maximum constant for a function, does that mean they know it lacks a minimum or maximum, or can they not prove whether it has a minimum or maximum.
I'd say the main difference is it involves adding, not subtracting, and I suspect there's a trivial minimum that is equal to 0, because I suspect the programming language can't subtract the Kolmogorov Complexity because it's defined to be the shortest program in a Turing-Complete language that outputs a given object, and the challenge is whether that minimum number of 0 actually exists, or whether the lower bound is higher than that for a programming language while still being Turing-Complete.
I think K-complexity is actually like that. For any given object you can define a Turing-complete language in which the empty program outputs that object. It could literally be "Python, except the empty program outputs this specific object".
So when you say that K-complexity is like the function that you describe earlier, does that mean that there's provably no minimum, or does that mean that you can't prove that the minimal constant for Kolmogorov Complexity exists?
It's still possible that I'm misunderstanding the question, but if it means what I think it means, then the answer is "provably no minimum".
I can accept this as an answer, though you'd have to show why Kolmogorov Complexity's additive constant lacks a minimum number more than you have in this comment thread.
However, "Python, except the empty program outputs this specific object" would probably be more complex than "Python", in most programming languages. So I wonder whether it would be possible to define objective complexity as eigenvector (not sure I am using the right word here) of relative complexities. As in: "simple" means "simple, when programmed in a simple language".
This is a question that I got from these posts, which essentially talked about how the constant issue impacts discussions of simplicity, and thus I wanted to ask a question about Kolmogorov Complexity:
https://www.lesswrong.com/posts/uWdAKyHZMfoxDHcCC/simplicity-arguments-for-scheming-section-4-3-of-scheming#What_is__simplicity__
https://joecarlsmith.com/2021/10/29/on-the-universal-distribution#iii-is-everything-equivalent-up-to-a-finite-fudge-factor
https://en.wikipedia.org/wiki/Kolmogorov_complexity#Invariance_theorem
What's the lowest additive constant that can be achieved by a programming language while still being Turing-Complete?
Or equivalently, what's the least overhead that you must have in order to describe an object like a computer program or string?
Bonus points for not only establishing a lower bound, but either finding a programming language that achieves the lower bound and showing the source, or actually creating a programming language that achieves the lower bound.
As a side question, I'd also like estimations of how much overhead/the additive constant is added to popular languages like Python, Java, C, C++, and Rust.