Jump to content

Self-modifying code

From Wikipedia, the free encyclopedia

This is an old revision of this page, as edited by 64.86.28.237 (talk) at 16:43, 15 April 2007 (Add COBOL reference ~~~~). The present address (URL) is a permanent link to this revision, which may differ significantly from the current revision.

In computer science, self-modifying code is code that alters its own instructions, whether or not it is on purpose, while it is executing.

Self-modifying code is quite straightforward to write when using assembly language (taking into account the CPU cache). It is also supported by some high level language interpreters such as SNOBOL4, the Lisp programming language, or the legendary ALTER verb in COBOL. It is more difficult to implement on compilers but compilers such as Clipper and Spitbol make a fair attempt at it, and COBOL almost encouraged it. Batch programming scripts often involve self-modifying code as well.

Usage of self-modifying code

Self-modifying code can be used for various purposes:

  1. Optimization of a state dependent loop.
  2. Runtime code generation, or specialization of an algorithm in runtime or loadtime (which is popular, for example, in the domain of real-time graphics).
  3. Altering of inlined state of an object, or simulating the high-level construction of closures.
  4. Patching of subroutine address calling, as done usually at load time of dynamic libraries. Whether this is regarded as 'self-modifying code' or not is a case of terminology.
  5. Evolutionary computing systems such as genetic programming.
  6. Hiding of code to prevent reverse engineering, as through use of a disassembler or debugger.
  7. Hiding of code to evade detection by virus/spyware scanning software and the like.
  8. Compression of code to be decompressed and executed at runtime, e.g. when memory or disk space is limited.
  9. Some very limited instruction sets leave no option but to use self-modifying code to achieve certain functionality. For example, a "One Instruction Set Computer" machine that uses only the subtract-and-branch-if-negative "instruction" cannot do an indirect copy (something like the equivalent of "*a = **b" in the C programming language) without using self-modifying code.

The second and third types are probably the kinds mostly used also in high-level languages, such as LISP.

Self-modifying code used to optimize a state-dependent loop

Pseudocode example:

repeat N times {
  if STATE is 1
   increase A by one
  else
   decrease A by one

  do something with A
}

Self-modifying code in this case would simply be a matter of rewriting the loop like this:

 repeat N times {

  increase A by one
  do something with A
 }
 
 when STATE has to switch {
    replace the opcode "increase" above with the opcode to decrease
 }

Note that 2-state replacement of the opcode can be easily written as 'xor var at address with the value "opcodeOf(Inc) xor opcodeOf(dec)"'.

Choosing this solution will have to depend of course on the value of 'N' and the frequency of state changing.

Attitudes towards self-modifying code

Some claim that use of self-modifying code is not recommended when a viable alternative exists, because such code can be difficult to understand and maintain.

Others simply view self-modifying code as something one would be doing while editing code (in the above example, replacing a line, or keyword), only done in run-time.

Self-modifying code was used in the early days of computers in order to save memory space, which was limited. It was also used to implement subroutine calls and returns when the instruction set only provided simple branching or skipping instructions to vary the control flow. This application is still relevant in certain ultra-RISC architectures, at least theoretically; see for example One instruction set computer. Donald Knuth's original MIX architecture also used self-modifying code to implement subroutine calls.

Self-modifying code used as camouflage

Self-modifying code was used to hide copy protection instructions in 1980s DOS based games. The floppy disk drive access instruction 'int 0x13' would not appear in the executable program's image but it would be written into the executable's memory image after the program started executing.

Self-modifying code is also sometimes used by programs that do not want to reveal their presence — such as computer viruses and some shellcodes. Viruses and shellcodes that use self-modifying code mostly do this in combination with polymorphic code. Polymorphic viruses are sometimes called primitive self-mutators. Modifying a piece of running code is also used in certain attacks, such as buffer overflows.

Operating systems and self-modifying code

Because of the security implications of self-modifying code, some operating systems go to lengths to rule it out. Recent versions of OpenBSD, for instance, have a feature known as W^X (for "write xor execute", meaning that a program can only write to or execute a given memory page, but not both). Versions of OpenBSD with W^X do not allow alteration of most memory pages which harbor executable code. Programs which depend upon rewriting their own machine code must use mmap to allocate pages with PROT_EXEC | PROT_WRITE protections.

Just-in-time compilers

Just in time compilers for Java, .NET and other programming languages often compile short blocks of byte-code or p-code into machine code suitable for the host processor and then immediately execute them.

Interaction of cache and self-modifying code

In some cases self-modifying code executes more slowly on modern processors. This is because a modern processor will usually try to keep blocks of code in its cache memory. Each time the program rewrites a part of itself, the rewritten part must be loaded into the cache again, which results in a slight delay.

The cache invalidation issue on modern processors usually means that self-modifying code would still be faster only when the modification will occur rarely, such as in the case of a state switching inside an inner loop. This consideration is not unique to processors with code cache, since on any processor rewriting the code never comes for free.

Most modern processors load the machine code before they execute it, which means that if an instruction that is too near the instruction pointer is modified, the processor will not notice, but instead execute the code as it was before it was modified. See Prefetch Input Queue (PIQ). PC processors have to handle self modifying code correctly for backwards compatibility reasons but they are far from efficient at doing so.

Example JavaScript self-modifying code

action = function(x) { return x+1; }

code = function() {
  for (i=1; i>0 && i<10; i=action(i)) {
    if (i==5) {
      action = function(x) { return x-1; }
      // if i=5 then modify code
    }
    document.write(i); // displays the results
  }
}

code()

would print: 1 2 3 4 5 4 3 2 1

Example NASM-syntax self-modifying x86-assembly algorithm that determines the size of the Prefetch Input Queue

 code_starts_here:
   xor cx, cx                  ; zero register cx
   xor ax, ax                  ; zero register ax
 
around:
   cmp ax, 1                   ; check if ax has been altered
   je found_size

   mov [nop_field+cx], 0x90    ; 0x90 = opcode "nop" (No OPeration)
   inc cx

   jmp short flush_queue
flush_queue:

   mov [nop_field+cx], 0x40    ; 0x40 = opcode "inc ax" (INCrease ax)

nop_field:
   nop times 256
   jmp around
found_size:

   ;    register cx now contains the size of the PIQ

What this code essentially does is change the execution flow, and determine by brute force how large the PIQ is. "How far away do I have to change the code in front of me for it to affect me?" If it is too near (it is already in the PIQ) the update will not have any effect. If it is far enough, the change of the code will affect the program and the program has then found the size of the processor's PIQ. If this code is being executed in protected mode, the operating system must not make any context switch, or else this program may return the wrong value.

Henry Massalin's Synthesis Kernel

The Synthesis kernel written by Dr. Henry Massalin as his PhD. thesis is commonly viewed to be the "mother of all self-modifying code." Massalin's tiny Unix kernel takes a structured, or even object oriented, approach to self-modifying code, where code is created for individual objects, like filehandles, and such code may be reoptimized later.

The Synthesis kernel is blindingly fast, but was written entirely in assembler. The resulting lack of portability has prevented Massalin's optimization ideas from being adopted by any production kernel. However, the structure of the techniques suggests that they could be captured by a higher level language, albeit one more complex than existing mid-level languages. Such a language and compiler could allow development of extremely fast operating systems and applications.

Paul Haeberli and Bruce Karsh have objected to the "marginalization" of self-modifying code, and optimization in general, in favor of reduced development costs; drawing a parallel to the "heavy religious atmosphere" which the Italian Futurist movement rebelled against.

See also