On Thu, 8 Feb 2007, Linus Torvalds wrote:
> No, making bitfields unsigned is actually usually a good idea. It allows
> you to often generate better code, and it actually tends to be what
> programmers _expect_. A lot of people seem to be surprised to hear that a
> one-bit bitfield actually often encodes -1/0, and not 0/1.
>
Your struct:
struct dummy {
int flag:1;
} a_variable;
should expect a_varible.flag to be signed, that's what the int says.
There is no special case here with regard to type. It's traditional K&R
that writing signed flag:1 here is redundant (K&R pg. 211). Now whether
that's what you wanted or not is irrelevant. As typical with C, it gives
you exactly what you asked for.
You're arguing for inconsistency in how bitfields are qualified based on
their size. If you declare a bitfield as int byte:8, then you're going
to expect this to behave exactly like a signed char, which it does.
struct {
int byte:8;
} __attribute__((packed)) b_variable;
b_variable.byte is identical to a signed char.
Just because a_variable.flag happens to be one bit, you cannot say that
programmers _expect_ it to be unsigned, or you would also expect
b_variable.byte to act as an unsigned char. signed is the default
behavior for all types besides char, so the behavior is appropriate based
on what most programmers would expect.
> So unsigned bitfields are not only traditional K&R, they are also usually
> _faster_ (which is probably why they are traditional K&R - along with
> allowing "char" to be unsigned by default). Don't knock them. It's much
> better to just remember that bitfields simply don't _have_ any standard
> sign unless you specify it explicitly, than saying "it should be signed
> because 'int' is signed".
>
Of course they're faster, it doesn't require the sign extension. But
you're adding additional semantics to the language by your interpretation
of what is expected by the programmer in the bitfield case as opposed to a
normal variable definition.
> I will actually argue that having signed bit-fields is almost always a
> bug, and that as a result you should _never_ use "int" at all. Especially
> as you might as well just write it as
>
> signed a:1;
>
> if you really want a signed bitfield.
>
How can signed bitfields almost always be a bug? I'm the programmer and I
want to store my data in a variable that I define, so I must follow the
twos-complement rule and allot a sign bit if I declare it as a signed
value. In C, you do this with "int"; otherwise, you use "unsigned".
> So I would reall yrecommend that you never use "int a:<bits>" AT ALL,
> because there really is never any good reason to do so. Do it as
>
> unsigned a:3;
> signed b:2;
>
> but never
>
> int c:4;
>
> because the latter really isn't sensible.
>
That _is_ sensible because anything you declare "int" is going to be
signed as far as gcc is concerned. "c" just happens to be 4-bits wide
instead of 32. But for everything else, it's the same as an int. I know
exactly what I'm getting by writing "int c:4", as would most programmers
who use bitfields to begin with.
David
-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [email protected]
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
[Index of Archives]
[Kernel Newbies]
[Netfilter]
[Bugtraq]
[Photo]
[Stuff]
[Gimp]
[Yosemite News]
[MIPS Linux]
[ARM Linux]
[Linux Security]
[Linux RAID]
[Video 4 Linux]
[Linux for the blind]
[Linux Resources]