Talk:bitch
This is the talk page for bitch.
Use this to discuss syntax, programs, or even to prove stuff about the language!
Note: For the sake of everyone's collective sanity, do not delete anything!
Even if nothing comes out of one of these discussions they are still useful to show thought processes and progress within certain problems.
If you don't like what you said or you changed your mind, strike it out (use <strike></strike>
or <del></del>
) and write underneath.
Contents
 1 Attempts at proving Turingcompleteness
 2 Alternative memory structures
 3 [Disproven] Conjecture: bitch cannot copy one bit from memory to another part of memory while retaining existing data
 4 Sketch: A RAM Machine
 5 Sketch: A Stack Machine
 6 Page content deletions
 7 Sketch: A Turing Machine
Attempts at proving Turingcompleteness
Since bitch is a newly created language and is fairly different to a lot of the languages usually used to prove Turingcompleteness, a proof is hard to come up with.
Below are the attempts currently ongoing. Feel free to join one or add a brand new one!
[In Progress] Compilation from Home Row to bitch by User:Helen
This section has been redone as of 16:43, 27 May 2019 (UTC). The previous version held nothing of use for either proving or showing the process behind proving the computational class of bitch.
This proof consists of two parts: a memory equivalence proof, and an operation equivalence proof.
In the memory equivalence proof, we will prove that Home Row's memory structure can be simulated in (and therefore is equivalent to) bitch's.
In the operation equivalence proof, we will prove that all Home Row's operations can be simulated as a whole using bitch's.
Any looping inside proof code will prevent looping in a program, due to nonnestable loops in both languages.
Since looping must remain possible as part of the equivalence, looping must not be present in proof code.
[Invalid] Memory equivalence
For an n
xm
Home Rowlike grid to be equivalent to a tape, we have to show that we can either:
 Simulate a grid inside a tape  every time we reach a cell whose index is a multiple of
grid width
(1indexed) and try to go forwards again, we must go backgrid width  1
cells
 every time we want to go down, we must go forwards grid width
cells
 Simulate a tape inside a grid  every time we want to go forwards in the tape, we must check whether the pointer is at the rightmost end of the grid, and if so, we wrap down to the leftmost end of the next row
For the sake of simplicity, neither of these has to be able to go back/up a cell.
This means the only way to backtrack is by wrapping back around at the right and bottom edges (for a grid) or at the end (for a tape).
Since it is more simple to simulate a tape in Home Row than it is to simulate a grid in bitch, this proof will show that it is possible to simulate a tape.
This means we must be able to check whether we are at the rightmost edge.
The simplest way to do this is by filling the rightmost column with zeros and not allowing any zeros inside cells (labelled c_{n}
):
c_{1 } c_{2 } c_{3 } c_{4 } 0 c_{5 } c_{6 } c_{7 } c_{8 } 0 c_{9 } c_{10} c_{11} c_{12} 0 c_{13} c_{14} c_{15} c_{16} 0 c_{17} c_{18} c_{19} c_{20} 0
We must then create a code snippet to prove we can move forwards.
For legibility and ease of understanding, two macrolike operations will be used:
l
 moves 1 grid space left, equivalent tof f f f
jl
is equivalent tojf jf jf jf
and not tojf f f f
.
u
 moves 1 grid space up, equivalent tod d d d
ju
is equivalent tojd jd jd jd
and not tojd d d d
.
Move forwards:
f
Move down if the pointer is on a zero:
d ju
Move right if the pointer is on a zero:
jl f
Combined:
f d jdjdjdjd jfjfjfjf f
We can navigate the Home Row grid as if it were a tape, with the limitation that the tape is only 20cells long.
However, this still proves that a grid memory space is equivalent to a tape space, since it shows that any n x m
grid is equivalent to at least an (n1)m
tape.
Overall size is irrelevant, as Turingcomplete actually requires and infinite tape size, and all languages are limited to a finite memory.
I don't believe that the translation of jl
works: jf jf jf jf
will get stuck when it hits a zero value. I believe the translation is also at odds with the requirement that the c_{i}
values are nonzero... Inte (talk) 12:59, 28 May 2019 (UTC)
 Both of those are intentional:
jl
is meant to get stuck on zero valuesc_{i}
values are meant to be limited to nonzero values
 However this is pretty clumsy and part of the reason I marked this as invalid.
 The new proof will hopefully be better, although a proof is pretty trivial and not really necessary but good for completion.  Helen (talk) 22:27, 29 May 2019 (UTC)
OTOH, since you're basing the memory on the cellbased storage, you can directly implement arbitrary permutations of c_{1} ... c_{nm}
, so the d
and f
operations can be implemented directly. Inte (talk) 12:59, 28 May 2019 (UTC)
 That's true the operations can most likely be directly implemented on a tape.
 I realised shortly after posting this proof that the memory structure I'm using for bitch is actually simply a cellbased RAM.
 For that reason, I invalidated the memory proof and I'm working on a new one.  Helen (talk) 22:27, 29 May 2019 (UTC)
Operation equivalence
In this proof, it is assumed that the random access cellbased alternative memory structure is being used.
There are some trivial equivalences (such as l
or j
), which we will deal with first.
However, there are also some more complex equivalences which will be dealt with individually and with explanations.
For the sake of simplicity, we will be using three macrolike operations:
Ln
 loads then
^{th} cell, equivalent to the code for loading the value of a cell.Sn
 sets then
^{th} cell, equivalent to the code for setting the value of a cell.
Trivial equivalences
Home Row  bitch 

j 
;

k at cell n 
Ln/&0Sn

l 
> or <

; 
.

Both f
and d
are not necessary to find an equivalence to, since we already showed that a cellRAM in bitch is equivalent to a grid in Home Row.
Addition and Subtraction
Addition (and subtraction) for a fixedsize cell is possible with definite looping and therefore can be unrolled to have no looping at all.
We remain with the same simple addition algorithm, instead we now write it for a fixed number of loops.
We must pick a number of loops such that we guarantee that p = oldP
at the end (or equivalently that q = 0
).
The bit width of the cells is enough to guarantee this.
p < some cell q < some cell p < p XOR q \ q < (p AND q) LS 1 \ ... > bitwidth times p < p XOR q / q < (p AND q) LS 1 /
We can use a pseudoextension of the cellRAM to temporarily store p
and q
.
We then perform the n
iterations and leave the result in the topmost accumulator.
Getting both variables:
This is the same as the code snippets from the code for performing bitwise operations.
[(a*n)&(1<<((ap)*n)1<<((ap1)*n))]((ap1)*n)&((1<<n)1)[((a+1)*n)]((a+1)*n)]((a+1)*n)&0[(a*n)&(1<<((aq)*n)1<<((aq1)*n))]((aq1)*n)&((1<<n)1)[((a+1)*n)[((a+2)*n)]n#0]((a+2)*n)&0
Storing the next p
value:
[(2*n)^]n]n&0
Storing the next u
value:
[(3*n)&(1<<(2*n)1)&]n[(n+1)&(1<<(2*n)1<<(n+1))]n&(1<<n1)[(3*n)&(1<<(4*n)1<<(2*n))](2*n)](2*n)&0
We then repeat those two code snippets another n1
times.
In the combined code, we'll paste this only once.
Moving the result into the accumulator:
[(2*n)]n&(1<<n1)
Combined:
[(a*n)&(1<<((ap)*n)1<<((ap1)*n))]((ap1)*n)&((1<<n)1)[((a+1)*n)]((a+1)*n)]((a+1)*n)&0[(a*n)&(1<<((aq)*n)1<<((aq1)*n))]((aq1)*n)&((1<<n)1)[((a+1)*n)[((a+2)*n)]n#0]((a+2)*n)&0 [(2*n)^]n]n&0[(3*n)&(1<<(2*n)1)&]n[(n+1)&(1<<(2*n)1<<(n+1))]n&(1<<n1)[(3*n)&(1<<(4*n)1<<(2*n))](2*n)](2*n)&0 ...n [(2*n)]n&(1<<n1)
Subtraction can be expressed as two additions  p + ~q + 1
.
Loading and negating q
:
[(a*n)&(1<<((ap)*n)1<<((ap1)*n))]((ap1)*n)&((1<<n)1)~]n~
Adding 1 to ~q
:
1]n [(2*n)^]n]n&0[(3*n)&(1<<(2*n)1)&]n[(n+1)&(1<<(2*n)1<<(n+1))]n&(1<<n1)[(3*n)&(1<<(4*n)1<<(2*n))](2*n)](2*n)&0 ...n [(2*n)]n&(1<<n1)]n
Loading p
:
[((a+1)*n)&(1<<((ap)*n)1<<((ap1)*n))]((ap1)*n)&((1<<n)1)]n
Summing p
and ~q + 1
:
[(2*n)^]n]n&0[(3*n)&(1<<(2*n)1)&]n[(n+1)&(1<<(2*n)1<<(n+1))]n&(1<<n1)[(3*n)&(1<<(4*n)1<<(2*n))](2*n)](2*n)&0 ...n
Moving the result into the accumulator:
[(2*n)]n&(1<<n1)
Combined:
[(a*n)&(1<<((ap)*n)1<<((ap1)*n))]((ap1)*n)&((1<<n)1)~]n~1]n [(2*n)^]n]n&0[(3*n)&(1<<(2*n)1)&]n[(n+1)&(1<<(2*n)1<<(n+1))]n&(1<<n1)[(3*n)&(1<<(4*n)1<<(2*n))](2*n)](2*n)&0 ...n [(2*n)]n&(1<<n1)]n[((a+1)*n)&(1<<((ap)*n)1<<((ap1)*n))]((ap1)*n)&((1<<n)1)]n[(2*n)^]n]n&0[(3*n)&(1<<(2*n)1)&]n[(n+1)&(1<<(2*n)1<<(n+1))]n&(1<<n1)[(3*n)&(1<<(4*n)1<<(2*n))](2*n)](2*n)&0 ...n [(2*n)]n&(1<<n1)
Conclusion
Since bitch can successfully emulate Home Row's operations and, their memory structures are equivalent, bitch must be able to solve at least as many problems as Home Row.
This means bitch is Turingcomplete.
I am pretty sure that brainfuck cannot compile to bitch, since brainfuck requires nested loops to be Turingcomplete. bitch can only nest an infinite loop once. I suggest that Home Row might result a more efficient compilation proof. A (talk) 11:08, 15 April 2019 (UTC)
Of course, I completely forgot about the loops. I had a look at Home Row and it seems to be all good but I didn't quite understand the loop definition. Is it similar to bitch's?  Helen (talk) 13:14, 15 April 2019 (UTC)
Oh I understand the loop now. It might be a problem since bitch's loops runs for the first time no matter what whereas Home Row's runs only if the current memory value is 0.  Helen (talk) 13:24, 15 April 2019 (UTC)
Anyway, the way of executing loops does not affect its Computational class. A (talk) 13:50, 15 April 2019 (UTC)
You are understanding the loops the wrong way. It says that the loop will run if the current memory spot is nonzero. A (talk) 14:55, 15 April 2019 (UTC)
Oh of course, my bad. But either way there is no reasonable way to replicate that in bitch in a general way. I'm glad it doesn't affect the computational class; phew  Helen (talk) 19:57, 15 April 2019 (UTC)
The compilation probably failed
Home Row requires unbounded memory to be Turingcomplete. However, only a limited 32bit version of incrementing and decrementing was implemented. So, the attempt had certainly failed. I could not find any other computational models useful for proving whether bitch is Turingcomplete. A (talk) 14:05, 13 May 2019 (UTC)
bitch's cellbased alternative storage is computationally equivalent to Home Row's.  Helen (talk) 17:18, 13 May 2019 (UTC)
 Sorry, but I believe you've missed an important point about Home Row: its 5^{2} cells are unbounded. That is how it manages to be Turingcomplete. So A has a point (with 32 replaced by the now arbitrary, but fixed, bit width) Inte (talk) 12:59, 28 May 2019 (UTC)
 bitch can simulate arbitrary sized cells. This is computationally the same as Home Row as no cell can be truly unbounded.
 Even if the cells are allowed to expand as much as necessary, eventually a RAM limit will be hit and the cell will be capped at an arbitrary size limit.
 bitch will be able to simulate a cell at that size limit as it too is only limited by the same factors.
 bitch can simulate arbitrary sized cells. This is computationally the same as Home Row as no cell can be truly unbounded.
 The fact remains that you're not showing Turingcompleteness. One can argue, and I would agree, that Turingcompleteness is a purely theoretical concept, and that in practice we're usually programming sufficiently large finite automata anyway. (Sometimes we're dealing with analogue signals, but those are noisy.) Note that just like Turing machines, bitch cannot be implemented on real hardware, because it has unbounded memory (both in the accumulator and in the storage). To show that bitch is TC, one needs to implement a Turingcomplete formalism (a Turing machine, or something equivalent to a Turing machine) in bitch.
 On the level of finite state machines, programming becomes a matter of compression: how big does a program have to be to operate on a certain number of states? A direct description of FSMs in terms of states and transitions is impractical for real computationsa program of size n can access O(n) states at best. With bitch, compression is possible: Cellbased storage (which you're using here) allows a program of size n to work on 2^{O(n)} states. The RAM construction boosts this to 2^{2O(n)}. Inte (talk) 13:39, 30 May 2019 (UTC)
 Yeah after thinking about it for a while, I realised I only really showed that it's equivalent to a boundedstorage machine with an arbitrarily large bound, which makes it "effectively TC" for all reasonable computation, but still doesn't prove that the language itself is Turingcomplete.
 Of course by "only really showed" I mean "will only really show" because I'm still working on finishing the new memory equivalence proof but once that's done I'll update this.
 I think I have an idea for addressing infinite memory in bitch, which would prove its Turingcompleteness, but I'll have to do that after this proof.  Helen (talk) 19:41, 30 May 2019 (UTC)
Terminology
You're attempting to show that every Home Row program can be translated to an equivalent bitch program, but for an equivalence proof between the two languages, you would need both directions, i.e., you'd also have to translate bitch programs to Home Row. What you have (if correct) is a reduction from a bounded cell variant of Home Row to bitch. Inte (talk) 12:59, 28 May 2019 (UTC)
Some thoughts
The provided implementation has a 63bit accumulator and bitstorage, which is pretty small, and Java has builtin bigints so you might want to use that. Also, according to this StackExchange answer, a 2stack Pushdown automaton, which might be emulated using the accumulator and bitstorage, is equivalent to a turing machine. So this could be a possible way of proving TCness. Although the lack of nested loops might be a problem. TuxCrafting (talk) 08:50, 17 April 2019 (UTC)
 There also seems to be some inconsistency with rapport to
>...<
: the spec says for<
that it "jumps to the nearest>
" but the implementation saves the position of a>
only when executed. So for example in>#0;><
, should the<
jump to the first or second>
?I will ask Helen; I have no idea.A (talk) 11:26, 17 April 2019 (UTC) Wait, where did you find that code snippet?
 I wrote it to show the problem. It's hypothetical, but having a welldefined spec is always a good thing. TuxCrafting (talk) 11:41, 17 April 2019 (UTC)
 I ran a similar program and found that it jumps to the first ">". You are right; the documentation indeed has to be modified.A (talk) 12:04, 17 April 2019 (UTC)
 I wrote it to show the problem. It's hypothetical, but having a welldefined spec is always a good thing. TuxCrafting (talk) 11:41, 17 April 2019 (UTC)
:I think that implementing a 2stack Pushdown automaton would be trivial, as 2 stacks already have the power of simulating a tape. It might be better if you directly simulate a tape. A (talk) 14:52, 29 April 2019 (UTC)
Spec problem fixed and the rest talked about in my user talk.  Helen (talk) 15:16, 17 April 2019 (UTC)
Another idea
It seems that we are having trouble implementing my incrementing/decrementing algorithm. That is why I made a improvement of the incrementing/decrementing algorithm.
This is what I typed in a Lua prompt. This improvement is inspired by this:
Lua 5.3.5 Copyright (C) 19942018 Lua.org, PUCRio > ~1 0 > ~1 2 >
~x is equal to x decremented, and ~x is equal to x incremented.
x is the negation of x. ~x is a bitwise NOT.
In case you question me about my algorithm, I will add more examples here:
> ~9 8 > ~9 10 > ~12 11 > ~12 13 >
These expressions are universal, which means that you can also run it in Java, Python, etc.
It requires a frontregister, which saves a pseudotop element of a number that changes dynamically when a new bit is used. Negating inverts the pseudotop element, and bitwise NOTting inverts all the bits from the pseudotop to the bottom. If there are 8 bits for each cell, 3 bits are sufficient for representing the frontregister. However, this results in recursive incrementing the frontregister every time a bit is used, which might be quite hard to do.
Inverting the bits after the pseudotop might be sufficient.
It is highly probable that I am misunderstanding these expressions.A (talk) 10:01, 29 April 2019 (UTC)
Bitch might be Turingincomplete
If we treated the accumulator as a counter, with [1
for increment and ]1
for decrement, we could treat it as a 1register Minsky machine. However, two counters are required for an incrementdecrement machine to be Turing complete, as onecounter Turing equivalence requires multiplication and division, which are impossible without a temporary.
Then, if we treat the accumulator+shift register as an unbounded bit tape, it still cannot be Turing complete, as we don't have the ability to nondestructively check only one bit of said tape, only checking if all the bits in the accumulator section are zero or at least one bit is non zero.
Maybe other models are possible, but it does seem like it's not Turingcomplete. TuxCrafting (talk) 15:33, 17 April 2019 (UTC)
 You can't independently check the value of a specific accumulator. Neither can you increment or decrement them, without making a loop, which would prevent backward state changes due to the nonexistence of nested loops or gotos. TuxCrafting (talk) 11:21, 18 April 2019 (UTC)
 In addition, I'd like to see an algorithm that can copy a bit in the accumulator to a different position, but *without* using loops, as using a loop in such an atomic operation would instantly break Turingcompleteness due to the necessity of a global program loop to apply state changes. TuxCrafting (talk) 11:28, 18 April 2019 (UTC)
 To clarify: what I mean is that bitch doesn't have any way of control flow other than the nonnestable
><
construct and conditional forward jumps. However, for Turing completeness, conditional backward jumps are also required. This can be emulated by putting most of the logic in an unconditional loop, but then it is impossible to build inner loops. TuxCrafting (talk) 11:44, 18 April 2019 (UTC)
 To clarify: what I mean is that bitch doesn't have any way of control flow other than the nonnestable
Nevermind, I didn't realize you could put a valid bitch instruction after a OR, which allows copying data from the storage. The article is quite unclear on this point. It might just be Turing complete, actually. TuxCrafting (talk) 21:11, 21 April 2019 (UTC)
Completely my fault, the ability to nest instructions like that is one of the most important parts of the language but it gets about two sentences on the main article.
I'm gonna rewrite most of the main article soon; I've been gathering feedback about things that generally aren't very clear.  Helen (talk) 22:40, 21 April 2019 (UTC)
[Unsuccessful] Compilation to Z3 derivative by User:Helen
This attempt contains programs that might be useful to prove the Turingcompleteness of bitch.
User:A  Basic compilation
This programming language looks like Konrad Zuse's Z3, which was already proven to be Turingcomplete by Raúl Rojas here.
First of all, Z3 is capable of executing an infinite loop. This is achieved in bitch as the > <
block.
Also, Z3 has math operation evaluation capabilities.
Z3 can implement 5 operations: addition, subtraction, multiplication, division, and squarerooting.
Even though these arithmetic operations have not been shown to be implementable in bitch, only the operations of zeroing a value and incrementing a value will be necessary for Turingcompleteness, as mentioned here.
Raúl Rojas shows that an instruction set with load, store, increment, zero and unconditional branching is Turingcomplete.
However, Load and Store have to be indirect. They should be nested for at least 2 indices.
Also, infinite memory (a requirement for Turingcompleteness) can be implemented in the Z3.
Finally, Z3 has a way to stop the program by using an invalid operation. That is .
in bitch.
User:Helen  Implementing arithmatic operations
This part was proven partially unneccesary, as A had found a command set that is easier to compile to bitch. On the other hand, these programs might make bitch easier to program in.
The conditional snippet below is easy to replicate in bitch:
if (z = i) then 0 else 1
We simply use XOR to check if a number z
is equal to some constant
i
like this:
#z^i;#1
If they are equal, the result of XORing them is automatically 0.
If they are not, the result is some nonzero number which we can detect (through the conditional ;
) and then set the accumulator to 1.
Mathematical operations are a bit harder to pull off.
I have managed to get an algorithm for addition.
Subtraction can be imitated through addition with negatives (due to algorithm overflowing and actually giving the correct number). It can't.
I haven't managed to get an algorithm for multiplication or division yet.
Addition:
Binary addition is very simple.
Ignoring carry overs, each digit behaves like this:
a
b
a+b
0
0
0
0
1
1
1
0
1
1
1
0
This is equivalent to XOR.
In order to get full addition behaviour though (rather than just "fairly close but not really" addition behaviour), we have to consider carry overs.
This is how carry overs behave:
a
b
a+b's carry
0
0
0
0
1
0
1
0
0
1
1
1
This is equivalent to AND.
The carry overs are put onto the digit one order higher though.
Therefore, we left shift each carry once.
This gives us our full (recursive) algorithm.
In pseudocode:
a < input
b < input
oldA < a+1
while (a != oldA) {
withoutCarries < a XOR b
carries < (a AND b) LS 1
oldA < a
a < withoutCarries
b < carries
}
This is perfectly doable in v4.7* bitch (takes 2 inputs):
\[16\]32>[32&4294967295#0^[16281474976710656&]32]15&131070[31&281470681743360]16&4294901760[16#0&4294967295[32&281474976710655]32;<[16/
User:Helen  Proving required operations
In order to be Turingcomplete, a language requires a very minimal instruction set (as shown by Raúl Rojas) and infinite tape (by definition).
The required instruction set consists of: load, store, increment, zero and unconditional branching.
If a language can do these and can simulate an infinite tape (or an arbitrarily sized tape), it is Turingcomplete.
In order to show these are true, we must envision the bitch accumulator as a tape.
Cells are delimited portions of the tape that are n
bits wide (n
being an arbitrary choice of integer width).
Addressing is signed, negative numbers meaning backwards in the tape and positive numbers meaning forwards in the tape.
Loading and storing values can't really be emulated in bitch but a parallel of setting and accessing cells can be made.
Setting cells
Setting a cell to a constant value c
can be done:
#c
However, this zeroes the rest of the accumulator which impends simulation of a tape.
Therefore, we opt for:
c
Given that the cell at position 0 is equal to 0, this will set load the value c
.
However this is dependant on the cell being 0, which is not desirable.
There are two options: zeroing the cell at position 0, or moving right/left (arbitrarily) until we find an empty cell.
I choose to zero the cell in order to be able to maintain a knowledge of where things are positioned in memory.
Furthermore, this doesn't impend one from storing the value at position 0 elsewhere and therefore it is the best option.
With c
being the desired constant value and n
being the bitwidth of the cells in the tape:
((1<<n)1)^((1<<n)1)c
For the sake of simplicity, brackets represent an expression which must be calculated and put in the source code.
If n
is 8, the code would become:
255^255c
Accessing cells
A value at positive position p
can be accessed through the following code:
](p*n)
A value at negative position q
can be accessed through the following code:
[(q*n)
Incrementing
Incrementing is simply the addition of 1.
It is also currently not implemented or working.
Incrementing can be done by checking each bit, starting from the rightmost bit going left, until we reach a 0 bit.
That 0 bit becomes 1 and all the bits lower than it become 0.
In order to check individual bits we must OR individual bits with 1 and check for equality with the previous version.
In pseudocode:
number < input
previous < number
n < 0
while(number  (1<<n) = previous) {
n < n+1
}
number < number  (1<<n)
number < number ^ ((1<<n)1)
The algorithm is currently a work in progress.
Zeroing a cell
This was done in the section on setting bits but will be restated here with some explanation.
Zeroing can be done by ANDing with a number that has the desired bits  the bits you want to zero  zeroed and all other bits as ones.
However, due to the arbitrary size of bitch's accumulator, we cannot rely on this (well we can but I choose not too since calculating differences of powers of two is more tedious than changing one digit).
Therefore, ORXORing seems to be most suitable.
We can OR to set the desired bits to 1, then XOR to "flip" them, effectively zeroing them.
If we want to zero the cell at positive position p
with cell size n
, we use the following:
]p((1<<n)1)^((1<<n)1)
If we want to zero the cell at negative position q
with cell size n
, we use the following:
[q((1<<n)1)^((1<<n)1)
Both of these will end with the tape positioned over the cell you just zeroed.
Unconditional Branching
bitch lacks in unconditional branching.
It is possible to infinitely repeat code but that is all you can do unconditionally.
>code<
Alternative memory structures
The accumulatorstorage combination may be quite limiting in certain circumstances.
In this case, some alternative memory structures are implemented below in bitch for ease.
Random access cellbased memory
A cellbased memory can be emulated in bitch.
All that is required is an arbitrary choice of bitwidth for the cells, and a size (in cells) of the currently allocated memory.
We will call the bitwidth of cells n
and the size of the memory a
.
In terms of structuring in bitch's accumulatorstorage conjunction, the memory will be in topmost storage (closest to the accumulator) and the accumulator will remain empty to be used for calculations.
Loading the value of a cell
Given a cell indexed p
(in a 0based index where the leftmost cell is 0 and the rightmost is a1
), we can set the accumulator to its value by the following steps.
Copying up the entire memory to the accumulator:
[(a*n)
ANDing the cell that we want:
&(1<<((ap)*n)1<<((ap1)*n))
Then shift copying it down to the bottom of the accumulator:
]((ap1)*n)&((1<<n)1)
Combined:
[(a*n)&(1<<((ap)*n)1<<((ap1)*n))]((ap1)*n)&((1<<n)1)
Setting the value of a cell
Given a cell indexed "p" (in a 0based index where the leftmost cell is 0 and the rightmost is a1
), we can transfer the value from the accumulator to the cell by the following steps.
Moving the accumulator value to the correct position to line up with p:
[((ap1)*n)&(1<<((ap)*n)1<<((ap1)*n))
Clearing the old value:
[(a*n)(1<<((ap)*n)1<<((ap1)*n))^(1<<((ap)*n)1<<((ap1)*n))
Copying down the new one:
](a*n)
Clearing the accumulator:
](a*n)&0
Combined:
[((ap1)*n)&(1<<((ap)*n)1<<((ap1)*n))[(a*n)(1<<((ap)*n)1<<((ap1)*n))^(1<<((ap)*n)1<<((ap1)*n))](a*n)](a*n)&0
Bitwise operations on two cells
Performing bitwise operations with two cell arguments is trivial now that we have the load and set.
We can use the ^^
combination to perform arbitrary argument operations without messing with the storage.
Given an operation x
, we can load two cells 0indexed p
and q
and then perform q x p
.
Loading cell p
:
[(a*n)&(1<<((ap)*n)1<<((ap1)*n))]((ap1)*n)&((1<<n)1)
Loading cell q
:
[((a+1)*n)]((a+1)*n)]((a+1)*n)&0[(a*n)&(1<<((aq)*n)1<<((aq1)*n))]((aq1)*n)&((1<<n)1)[((a+1)*n)[((a+2)*n)]n#0]((a+2)*n)&0[(2*n)
Performing binary operation x
:
^^x]n]n&0[n
Combined:
[(a*n)&(1<<((ap)*n)1<<((ap1)*n))]((ap1)*n)&((1<<n)1)[((a+1)*n)]((a+1)*n)]((a+1)*n)&0[(a*n)&(1<<((aq)*n)1<<((aq1)*n))]((aq1)*n)&((1<<n)1)[((a+1)*n)[((a+2)*n)]n#0]((a+2)*n)&0[(2*n)^^x]n]n&0[n
[Disproven] Conjecture: bitch cannot copy one bit from memory to another part of memory while retaining existing data
 edit: The specific memory copy claim/challenge of this conjecture has been successfully disproven, multiple times over. The more general 'Not TC' part still seems valid, but I'm changing the section heading so it's clear what the 'Disproven' relates to. Section still has value because it prompted useful progress in determining computational class Salpynx (talk) 23:42, 20 May 2019 (UTC)
I was following TuxCrafting's (now struckthrough) arguments against TC above, and I don't think the ability to chain instructions helps that much (because they can't be grouped), and the challenge "to see an algorithm that can copy a bit in the accumulator to a different position" is still open.
There are multiple ways to formulate this as a challenge, but this is the simplest I could come up with:
Create a bitch program that can take an already initialised 3 bit accumulator abc
and convert it programatically to abb
:
abc => abb
examples:
011 => 011 (3 => 3)
101 => 100 (5 => 4)
110 => 111 (6 => 7)
(As I read it) There is no way to group or order the evaluation of chained operations, and no temporary scratch space to use different parts of the accumulator for a single operation. The accumulator cannot be practically masked in such a way that one subpart of it can interact with another and restored, without losing subparts not involved in the operation.
Here is the full mapping of expected results:
0 > 0
1 > 0
2 > 3
3 > 3
4 > 4
5 > 4
6 > 7
7 > 7
A simple algorithm to do this for three bits is:
(abc & 0b110)  ((abc & 0b010) >> 1)
This requires grouping and two separate logic operations on the accumulator (abc) as sub steps. Since bitch modifies the accumulator in place, no amount of instruction chaining will help. The logic operations are only useful when combining the accumulator with number literals or user input.
Conjecture:
 A machine with two independent stacks should be able to copy a value from one stack to the other. A machine that cannot copy a single bit in isolation from anywhere in memory to anywhere else, regardless of the actual datastructures involved, has no hope of simulating two stacks.
 There is no way to create a generalised algorithm to perform simple bit copy in bitch, therefore the language cannot simulate two independent storage stacks or accumulators.
 It may be possible to produce the results above in bitch by creating effectively a lookup table mapping the input to desired output, but that proves lack of Turing completeness if the only way to produce arbitrary results is to create lookup tables mapping input to output.
 The ability to copy data from one memory location to another with destroying all other data is required equally for any TC system involving multiple stacks, multiple counter registers, and equally to cyclic tag systems and cellular automata.
If I have somehow missed something in the spec, proving this false does not necessarily mean the language is TC. If there is a clever way to do the copy, I imagine it will require effectively placing an upper bound on the accumulator, and multiple unbounded stacks or accumulators cannot be simulated from a bounded one.
I believe this makes bitch a PDA, since one end of the accumulator can be used as a single stack, and transitions decision can be made based on that.
This talk page is getting hard to follow with many sections having questionable statements without attribution. Hopefully this bit copy challenge will be easier to address than a full TC proof. Salpynx (talk) 23:48, 7 May 2019 (UTC)
 I didn't notice the implemented memory structure was supposed to be bitch because it has parentheses
()
to group operators (which AFAICT bitch lacks, and is the major hurdle to completing this challenge in pure bitch) and standard bitshift operators <<
mixed in with bitch syntax. I am guessing that the intent was to precalculate all those values and substitute them into the bitch source as number literals? This is what I have noticed about all the examples given, and what this challenge is trying to prove: that binary operations cannot be done on two independent memory locations. It is easy to perform an operation with memory + number literal, but if you have to compute it all before hand and hardcode it into bitch, it's not bitch doing the computation, at least not in a way that is TC. I tried substituting values into that code and couldn't get it to work, can you? The form for this challenge will be something like \[]/
where []
is replaced with a bitch algorithm of any size which copies the bit. If it is easy to do, then I have clearly misunderstood something about how the storage behaves and an example would clear up the confusion. Salpynx (talk) 12:08, 8 May 2019 (UTC)
 In theory you can do any terminating amount of precomputation "by hand" (which really means as part of compiling into your language). (From Oerjan's user talk). A (talk) 04:02, 9 May 2019 (UTC)
 Copying a bit of a 3 bit accumulator can be achieved using the follwing piece of code:
&6]3[2&1[3]3&7
(which also keeps the storage unmodified). One key point here is that chained instructions work on a fresh copy of the state (both accumulator and storage), so [2
and similar instructions leave the storage unmodified.
 More generally, any map with finite domain (old accumulator value to new accumulator value) can be implemented, as for example the ROT13 program on the main page. I don't see how to copy a bit in an unbounded accumulator though, so none of this really brings us any closer to TCness. This should be enough to show that bitch can simulate arbitrary PDAs though. For processing input, note that one should be able to read from input without destroying the storage using
&0\
, but the Java interpreter doesn't really support this because each instance of Program creates its own new Scanner that reads a whole line of input... Inte (talk) 05:43, 9 May 2019 (UTC)
 Congratulations! This is a valid solution to the challenge as I stated it. I was surprised it works, but I can now see how, which is a welcome outcome. I think knowing the conditions under which this is possible is helpful too. Like I guessed it places bounds on the storage, and proves the language is at least computationally equivalent to a PDA. Perhaps there is a way to break though this limitation, but it makes the next steps clearer, and provides a clear example of a useful programming technique. Salpynx (talk) 00:03, 10 May 2019 (UTC)
This conjecture is disproven by the old alternative memory structures section as I understand it.
The variables there are simply the number of cells, the size of the cells and the cell to access.
The number of cells and size of cells is simply for customizability and would be fixed by the programmer to a value of their choosing for the entire program (e.g.: 20 and 8).
The position variable is simply the position at which the cell is that you want to access.
The expressions given are simply the values needed to move and zero parts of the accumulator as necessary, they work with any values.
Binary operations on two different locations are very possible and they are the only reason for chaining operators.
The code snippet &[8
will AND two different values given that they are both 8 bits wide and in the accumulator together, one on top of the other.
The tape memory structure doesn't rely on precomputing, it allows for different sized tapes and different sized cells for customizability.
The values for the tape code are no less arbitrary than the value 8 in the above code.  Helen (talk) 19:23, 9 May 2019 (UTC)
 I seem to have reinvented parts of the "alternative memory structures" section below for GET and PUT, though as far as I can see the main contribution, random access, is new. I don't know why A deleted that section... Inte (talk) 21:10, 9 May 2019 (UTC)
 I tried to use your cell based memory structure early on but couldn't get it to work, here is a worked example of what I get:
 With 3 bit memory, each cell is 1 bit, and the task is to copy cell 1 to the accumulator (which is slightly different from my challenge, but probably equivalent given some more steps), my abc is 012 in this structure.
 a = 3 (memory size)
 p = 1 (target cell, 0indexed)
 n = 1 (cell width)
 Loading a value from a cell formula, substituting using the values above:
[3&4]1
 I understand that this requires the memory to already be in the storage section, so if I preload the storage to 0b010 using
#2]3
, I would expect to get 1 in the accumulator by running #2]3[3&4]1
, but that results in 0 in the accumulator and 2 in storage. Trying it with storage initialised to 0b111 (7) results in 0b110 (6) in the accumulator and 0b111 (7) in storage. Perhaps the formulas don't work with a cell width of one, but I don't know under which conditions it does work. I don't think that section should be deleted, it's possible am missing something about how to use the structure properly, or that there is a tiny bug in one part of it that could be identified and fixed as a result of using it for a specific task. This challenge was an attempt to get clarity on how to use the theorised techniques to do a specific task that did not seem obviously trivial to me. I would love to see an working example of your cellbased memory structure applied to the task above, I did try but couldn't make it work. The addition example (I think that is yours too?) was great as I could easily verify it did exactly what it claimed, and then spend time trying to understand how. Salpynx (talk) 00:20, 10 May 2019 (UTC)
 Yeah sorry for the lack of any examples.
 Given your above values:
 a = 3 (memory size)
 n = 1 (cell width)
 p = 1 (target cell, 0indexed)
 The following code snippet stores and retrieves 1 from the 2nd position (2 line <pre> tag meant I couldn't indent; anyone who can format well is free to do this over):
#1[1&2[32^2]3]3&0[3&2]1&1/
 This is made up of two main segments:
 The storage segment:
[1&2[32^2]3]3&0
 The retrieval segment:
[3&2]1&1
 The solution to the original problem (directly copied and evaluated) would be:
]3[3&2]1&1[0&1[31^1]3]3&0[3
 Thanks Helen, that example does indeed fulfil my challenge using your cell based memory structure, congratulations! I see now how it works, that the memory structure is good, and that you had figured out usable memory before I proposed my challenge. Apologies for doubting you!
 Bitshift operator precedence seems to come after arithmetic operators in most programming languages (C, python, js), so when I used that formula, copied directly from Bitch#Loading_the_value_of_a_cell section,
1<<21<<1 = 4
which is what I tried for "ANDing the cell that we want" is different from (1<<2)(1<<1) = 2
, which looks like what you intended in the example. That threw me early and I stopped digging deeper. If you don't mind, I'll try to follow the rest of the example and update your Bitch#Loading_the_value_of_a_cell section if I can make it clearer? Salpynx (talk) 23:26, 20 May 2019 (UTC)
 Yeah, go ahead! That's perfectly fine by me.
 I always mentally replace
1<<
with 2^
so order of operations is a bit messed up for me on that sorry.
Sketch: A RAM Machine
We sketch a bitch implementation of a machine with a 32 bit accumulator A (stored in bits 0..31 of the accumulator),
a 32 bit extended accumulator X (bits 32..63), optional quickaccess registers, and up to 2^26  1 individually addressable words of RAM.
So even if bitch is not TC, it still appears to be a capable programming language. Inte (talk) 17:41, 9 May 2019 (UTC)
 I've used this machinery to make a finite memory (so nonTC) Brainfuck interpreter; see https://github.com/inte/bits/blob/master/examples/brainfuck.pp and the resulting https://github.com/inte/bits/blob/master/examples/brainfuck (not thoroughly tested, may have bugs) Inte (talk) 22:47, 16 May 2019 (UTC)
 In case other users get confused of how that preprocessor works, I clarify that the preprocessor takes input from the console and outputs the resulting bitch program. A (talk) 12:10, 20 May 2019 (UTC)
How is it supposed to work? (I have inputted a brainfuck program and could not get it to work...) Does it take a program from input or whatever?
 I will be testing your implementation using
,[.[>++<]>+]
, which makes sure that a brainfuck interpreter supports all commands, nested loops, and unbounded memory. Testing halting is trivial...
 I have an idea: you could extend the implementation to be unbounded by supplying tape extending operation, if this machinery supports extending the memory.
 First of all, extending memory after initialization is hard, perhaps even impossible, because the design is based on each memory cell containing its own address. (Without that stored address, `poke` doesn't work.) Shifting the existing memory would invalidate the addresses stored in that memory. Oh and I wouldn't know where to even begin doing such an extension without a loop.
 Note also that addresses are bounded. The current implementation uses 16 bit words, both for data and the addresses. Addresses need to be divisible by 32, so at most 2^11  1 = 2047 addressable memory cells are supported, and that's what the interpreter provides (minus memory used for stack, program, and registers).
 For the particular application as a Brainfuck interpreter, one could change the memory cell layout to use, for example, a 24 bit address and 8 bit data, increasing the address space to 2^19  1 = 512287 cells, but that's a nontrivial effort. An easier approach would be to set the word size to 32 bits (BB=5), with up to 2^27 addressable memory cells. Inte (talk) 14:56, 17 May 2019 (UTC)
 Wow, the macro replacements can act as variables!!! See this program(which seems to be an abuse of this preprocessor):
A=1,
,A,
A=2,
,A,
 This program prints
12
, which is 1 and 2 combined. This can demonstrate that the word sized can be set dynamically in the preliminary stage; however, there is currently no way to implement unbounded cells using this way.
This will not work if you want to extend the cell length though:
CNT=4,
BB=CNT,
CNT={BB+1},
BB=CNT,
,BB,
A (talk) 13:04, 22 May 2019 (UTC)
A summary of the construction in this section has found its way to the bitch page (last week). Is there anything left here that is worth preserving? (I mean the construction itself, not the discussion thereof.) Inte (talk) 08:09, 27 May 2019 (UTC)
Memory
The storage is used for memory. Addresses are 32 bits, the lowest 6 of
which are always 0. Each word occupies 64 bits in memory that store the
address of the word and the word itself value.
The first few addresses may be used as additional registers instead,
at the programmer's discretion.
Operations
None of these operations use > or <, so we can use them in loops.
The operations are not clean for the ; conditional in the sense that
the accumulator may become 0 while an instruction is executed.
It should not be too difficult to accomplish that, with some extra care.
(The motivation would be that operations that are clean for ; can be
used as part of a state machine that manages the control flow.)
 load immediate n (clears X):
IMM n := ^^n
 32 bit decrement (assumes X = 0):
DEC :=
]32^1
[32^]32]32&[32
[31^]31]31&[31
...
[2^]2]2&[2
[1^]1]1&[1
&0[32
 32 bit increment (assumes X = 0):
INC :=
~]32&0[32
DEC32
~]32&0[32
 load a register (assumes X = 0, n divisible by 32):
GET n := ^^[n]32&0[32
 store a register (assumes X = 0, n divisible by 32):
PUT n := [n&4294967296]n]n
 read a word at address in accumulator:
# comment format: # accumulator  reverse storage
PEEK := # A  ... A W ...
^^[^32 # A ... A W  ... A W ...
]32&0[32 # W  ... A W ...
 store word at address (<2^32  8) in accumulator, with value given in upper 32 bit of (extended) accumulator:
# comment format: # accumulator  reverse storage
POKE := # X A  ... A W (A+64) ...
[&4294967295 # X A ... A  W (A+64) ...
[32 # X A ... A W  (A+64) ...
&4294967296 # X A ... A 0  (A+64) ...
^^[32 # X A ... A 0 (A+64)  (A+64) ...
^]&4294967295 # X A ... A X ??  (A+64) ...
^^]32 # X A ... A X  (A+64) ...
]32 # X A ... A  X (A+64) ...
]&4294967295 # X A  ... A X (A+64) ...
 load extended part of accumulator (n divisible by 32) for stores:
GETX n :=
]32
GET (n+32)
[32
 clear extended accumulator:
CLEARX := &4294967295
 arithmetic with extended accumulator:
XOR := ^]32
AND := &]32
OR := ]32
NOT := ]32~]32~[64
 initialize memory (only needed once, at the start of the program, so using > < ; is okay):
INIT :=
>DEC&64]64[64;<[64
 Bitwise right shift accumulator A:
RSH n := ^^]n
 Bitwise left shift accumulator A:
LSH n := ^^[n
 Swap the two accumulators:
XCHG := ^]32^[32^]32
 Here are some extra instructions I find in Assembly language that might improve the usability of the instruction set(some are very trivial):
 I have a question about the IMM instruction. Couldn't you directly use
^x
(which works) instead of ^^x
? A (talk) 10:19, 10 May 2019 (UTC)
 The point is that
^^n
sets the accumulator to n independently of the previous value of the accumulator. You could achieve the same effect with &0^n
, but that would be one more character. Inte (talk) 14:06, 10 May 2019 (UTC)
I think I don't understand the [n^]n]n&[n
operations in the decrement implementation. I believe it does carrying, but I can't understand how it achives it. (this comment by A at 14:49, 10 May 2019 UTC UTC; please sign your comments with ~~~~)
Sketch: A Stack Machine
We sketch another machine with an accumulator (with 64 bits) and a stack(preserved in the storage).
This machine simulation in bitch is created merely to extend the power of bitch programs.
Note: If mixing this stack machine with the RAM machine above is neccesary, the two accumulators must be preserved somewhere inside the storage before using the commands.A (talk) 11:54, 13 May 2019 (UTC)
Memory
The memory is just an accumulator (stored in the bitch accumulator) and an unbounded stack (stored in the accumulator storage).
Operations
All of the commands do not use loops.
 Duplicate the top of the stack
DUP := ^^[32]32
 Push a number onto the stack
PUSH x := ^^x]32
 Pop a number
POP := [32^^0
 Pop a number, outputting it
POPOUT := [32/^^0
 Swap the top 2 elements of the stack
SWP := [64^]32^[32^]32]64
 Push input
PUSHI := &0\]32
 Increment top of stack
INC :=
[32~]32&0[32
DEC
~]32&0
 Decrement top of stack
DEC :=
^1
[32^]32]32&[32
[31^]31]31&[31
...
[2^]2]2&[2
[1^]1]1&[1
&0
 Check whether two topofstacks are equivalent
CHK := ^^[32^[64]32
 Check if top of stack > the second top of stack
 Check if top of stack < the second top of stack
 Implementation of +, , *, /, modulo, etc (impossible? *, /, and modulo possibly requires nested loops)
 Reverse the stack (impossible?)
Page content deletions
I'd like to point out that with MediaWiki's awful default diff, there's hardly anything worse for following talk page changes than deleting content. Also deleting other people's talk page contributions, or even your own after others have responded to them, is generally considered inappropriate, except on your own personal talk page.
We generally approximate Wikipedia policy here when we don't have a specific different tradition, so I'd sum up my understanding of what is done there: The recommended way to delete content from talk pages that are too big is to only move full sections to an archive subpage (like Talk:Bitch/Archive 1), and only when there has been no discussion in that section for a reasonably long time (which might depend on how fast discussion is happening).
Sketch: A Turing Machine
We can implement a Turing machine on a binary alphabet. The main challenge here is to implement bitwise boolean operations that keep (most of) the accumulator intact. It turns out that this isn't too hard if one pads the data with sufficiently many alwayszero bits. Inte (talk) 20:53, 23 May 2019 (UTC)
 Is the state represented using 1 bit? Also, what does k and q mean? (You don't have to explain the meaning of
k:l
and r:q
; I know them.)A (talk) 10:19, 24 May 2019 (UTC)
 The state is represented by as many bits as needed for the given Turing machine (the second s in ss indicates a plural; this convention comes from Haskell...). k is an adjacent letter to l and q is adjacent to r; I went down (in the alphabet Inte (talk) 12:59, 25 May 2019 (UTC)) rather than up to avoid confusion with s. Inte (talk) 12:00, 24 May 2019 (UTC)
 I don't know how that can "avoid confusion". Hmm, l is directly adjacent to s, which might arise confusion: l could be considered part of the ss. Also, you wrote
I went down rather than up to avoid confusion with s.
; what does I
mean? I did not find I
anywhere in this sketch.
The contents of this section has moved here. Inte (talk) 07:43, 27 May 2019 (UTC)