I apologize in advance for that I give way too many details, but I'm about to contradict people.
About K(x)≤K′(x)+cK(x)≤K′(x)+c
The fact that K1(x)≤K2(x)+cK1(x)≤K2(x)+c usually comes from an interpreter of the description language #2 into the description language #1 and not from a translation from programs of #2 into programs of #1.
For example KC(x)≤KPython(x)+cpy2cKC(x)≤KPython(x)+cpy2c and you get this inequality as simply as this:
void py_run(char * s) {
// code of your Python interpreter
}
int main(void) {
py_run("Put here your Python program of size Kpython(x)");
}
Then your constant cpy2ccpy2c will be something like 528+490240688528+490240688 where 528528 is the number of bits for this code and 490240688490240688 bits is the size the official Python interpreter written in C. Of course you only need to interpret what is possible in your description language for Python so you can do better than 69 MB :-)
What is important is that you can write your Python program linearly in your C code. For example, a language where you need to put "BANANA" between every character is not a very good description program and the property is then false. (But if the description language authorizes you to write data in a separate file or in a block, this problem disappear)
Why your K1(x)=qK1(x)=q is flawed
The problem with your definition of K1K1 is that you may need more than qq bits to describe a Turing machine with qq states because you need to encode transitions.
So no K1K1 and K2K2 are probably not equivalent, but that's mainly K1K1's fault. I think that we can prove that for all a>0a>0 there is a caca such that K1(x)≤a|x|+caK1(x)≤a|x|+ca. Of course any a<1a<1 is enough to disprove the fact that K1K1 is not a valid function, since it would mean that we can encode more all 2n2n possible strings of length nn into an+caan+ca bits.
But the size is an incredibly tight bound when building Turing machines. The idea is that in a block of bb states there are b2bb2b ways to find transitions for each state and that's better than the usual 2b2b ways you can fill bb bits. Then you can store in each block log2blog2b bits of information. (not 2log2b2log2b because you have to go in and out the block one way or another)
So yeah... With blocks of size 21/a21/a you could probably prove K1(x)≤a|x|+caK1(x)≤a|x|+ca. But I already written way too much about why the number of states is not a valid Kolmogorov complexity function. If you want me to elaborate, I will.
Now about K2K2
The naive descriptive language corresponds roughly to K2(x)=q⋅2⋅(log2q+2)K2(x)=q⋅2⋅(log2q+2) (i.e. log2qlog2q for each next state and details about write and termination).
As you seem to be, I'm convinced that a better/cheater way would be to authorize to encode "data" into Turing machine, maybe by adding a binary tag in the description language that says whether if a state is a data state (that just writes a bit and go to the next state) or if it does something else. That way you could store one bit of your xx in one bit of your descriptive language.
However if you keep the same K2K2 you could use the same technique I used in the previous part to save a few bits, but I seem to be stuck at K2(x)≤a|x|log|x|+cK2(x)≤a|x|log|x|+c (for any a>0a>0) .. maybe less than log|x|log|x|, even, but obtaining O(|x|) seems hard. (And I expect it should be |x|, not even O(|x|).)