Hi Guys , was wondering how PBP handles bits var defines , relative to byte defines. for use of memory code space use.
when i have a need of many flags i have generally defined a byte , and then allocated each bit of that byte to the var flag names as required ,
where when i need only 1 bit i define as a bit onlyCode:bit_byte1 var byte bit0 var bit_byte1.0 bit1 var bit_byte1.1 bit2 var bit_byte1.2 bit3 var bit_byte1.3 bit4 var bit_byte1.4 bit5 var bit_byte1.5 bit6 var bit_byte1.6 bit7 var bit_byte1.7 ' need only 1 bit then mybit var Bit
this is done in the assumption that when is compiled the memeory use will be less than if i had defined a byte
does this hold true ?




Bookmarks