Tuesday, April 16, 2024
HomeSoftware EngineeringEvaluating the Efficiency of Hashing Strategies for Comparable Perform Detection

Evaluating the Efficiency of Hashing Strategies for Comparable Perform Detection


Think about that you simply reverse engineered a chunk of malware in painstaking element solely to seek out that the malware creator created a barely modified model of the malware the subsequent day. You would not wish to redo all of your onerous work. One solution to keep away from beginning over from scratch is to make use of code comparability methods to attempt to determine pairs of capabilities within the previous and new model which might be “the identical.” (I’ve used quotes as a result of “similar” on this occasion is a little bit of a nebulous idea, as we’ll see).

A number of instruments may also help in such conditions. A highly regarded industrial device is zynamics’ BinDiff, which is now owned by Google and is free. The SEI’s Pharos binary evaluation framework additionally features a code comparability utility referred to as fn2hash. On this SEI Weblog publish, the primary in a two-part sequence on hashing, I first clarify the challenges of code comparability and current our answer to the issue. I then introduce fuzzy hashing, a brand new kind of hashing algorithm that may carry out inexact matching. Lastly, I examine the efficiency of fn2hash and a fuzzy hashing algorithm utilizing plenty of experiments.

Background

Actual Hashing

fn2hash employs a number of forms of hashing. Probably the most generally used is named place impartial code (PIC) hashing. To see why PIC hashing is necessary, we’ll begin by a naive precursor to PIC hashing: merely hashing the instruction bytes of a perform. We’ll name this precise hashing.

For instance, I compiled this easy program oo.cpp with g++. Determine 1 exhibits the start of the meeting code for the perform myfunc (full code):

Figure 1- Assembly code and bytes from oo.gcc

Determine 1: Meeting Code and Bytes from oo.gcc

Actual Bytes

Within the first highlighted line of Determine 1, you may see that the primary instruction is a push r14, which is encoded by the hexadecimal instruction bytes 41 56. If we acquire the encoded instruction bytes for each instruction within the perform, we get the next (Determine 2):

Figure 2-Exact bytes in oo.gcc

Determine 2: Actual Bytes in oo.gcc

We name this sequence the precise bytes of the perform. We are able to hash these bytes to get an precise hash, 62CE2E852A685A8971AF291244A1283A.

Shortcomings of Actual Hashing

The highlighted name at deal with 0x401210 is a relative name, which signifies that the goal is specified as an offset from the present instruction (technically, the subsequent instruction). For those who have a look at the instruction bytes for this instruction, it consists of the bytes bb fe ff ff, which is 0xfffffebb in little endian type;. Interpreted as a signed integer worth, that is -325. If we take the deal with of the subsequent instruction (0x401210 + 5 == 0x401215) after which add -325 to it, we get 0x4010d0, which is the deal with of operator new, the goal of the decision. Now we all know that bb fe ff ff is an offset from the subsequent instruction. Such offsets are referred to as relative offsets as a result of they’re relative to the deal with of the subsequent instruction.

I created a barely modified program (oo2.gcc) by including an empty, unused perform earlier than myfunc (Determine 3). Yow will discover the disassembly of myfunc for this executable right here.

If we take the precise hash of myfunc on this new executable, we get 05718F65D9AA5176682C6C2D5404CA8D. You’ll discover that is completely different from the hash for myfunc within the first executable, 62CE2E852A685A8971AF291244A1283A. What occurred? Let us take a look at the disassembly.

Figure 3 - Assembly code and bytes from oo2.gcc

Determine 3: Meeting Code and Bytes from oo2.gcc

Discover that myfunc moved from 0x401200 to 0x401210, which additionally moved the deal with of the decision instruction from 0x401210 to 0x401220. The decision goal is specified as an offset from the (subsequent) instruction’s deal with, which modified by 0x10 == 16, so the offset bytes for the decision modified from bb fe ff ff (-325) to ab fe ff ff (-341 == -325 – 16). These adjustments modify the precise bytes as proven in Determine 4.

Figure 4 - Exact bytes in 002.gcc

Determine 4: Actual Bytes in oo2.gcc

Determine 5 presents a visible comparability. Purple represents bytes which might be solely in oo.gcc, and inexperienced represents bytes in oo2.gcc. The variations are small as a result of the offset is just altering by 0x10, however this is sufficient to break precise hashing.

Figure 5- Difference between exact bytes in oo.gcc and oo2.gcc

Determine 5: Distinction Between Actual Bytes in oo.gcc and oo2.gcc

PIC Hashing

The aim of PIC hashing is to compute a hash or signature of code in a method that preserves the hash when relocating the code. This requirement is necessary as a result of, as we simply noticed, modifying a program usually ends in small adjustments to addresses and offsets, and we do not need these adjustments to change the hash. The instinct behind PIC hashing is very straight-forward: Establish offsets and addresses which might be more likely to change if this system is recompiled, similar to bb fe ff ff, and easily set them to zero earlier than hashing the bytes. That method, if they modify as a result of the perform is relocated, the perform’s PIC hash will not change.

The next visible diff (Determine 6) exhibits the variations between the precise bytes and the PIC bytes on myfunc in oo.gcc. Purple represents bytes which might be solely within the PIC bytes, and inexperienced represents the precise bytes. As anticipated, the primary change we are able to see is the byte sequence bb fe ff ff, which is modified to zeros.

Figure 6- Byte Difference Between PIC Bytes (Red) and Exact Bytes (Green)

Determine 6: Byte Distinction Between PIC Bytes (Purple) and Actual Bytes (Inexperienced)

If we hash the PIC bytes, we get the PIC hash EA4256ECB85EDCF3F1515EACFA734E17. And, as we might hope, we get the similar PIC hash for myfunc within the barely modified oo2.gcc.

Evaluating the Accuracy of PIC Hashing

The first motivation behind PIC hashing is to detect an identical code that’s moved to a unique location. However what if two items of code are compiled with completely different compilers or completely different compiler flags? What if two capabilities are very related, however one has a line of code eliminated? These adjustments would modify the non-offset bytes which might be used within the PIC hash, so it might change the PIC hash of the code. Since we all know that PIC hashing is not going to at all times work, on this part we focus on methods to measure the efficiency of PIC hashing and examine it to different code comparability methods.

Earlier than we are able to outline the accuracy of any code comparability method, we’d like some floor reality that tells us which capabilities are equal. For this weblog publish, we use compiler debug symbols to map perform addresses to their names. Doing so gives us with a floor reality set of capabilities and their names. Additionally, for the needs of this weblog publish, we assume that if two capabilities have the identical identify, they’re “the identical.” (This, clearly, is just not true typically!)

Confusion Matrices

So, for example we’ve got two related executables, and we wish to consider how properly PIC hashing can determine equal capabilities throughout each executables. We begin by contemplating all attainable pairs of capabilities, the place every pair comprises a perform from every executable. This method is named the Cartesian product between the capabilities within the first executable and the capabilities within the second executable. For every perform pair, we use the bottom reality to find out whether or not the capabilities are the identical by seeing if they’ve the identical identify. We then use PIC hashing to foretell whether or not the capabilities are the identical by computing their hashes and seeing if they’re an identical. There are two outcomes for every willpower, so there are 4 prospects in complete:

  • True optimistic (TP): PIC hashing accurately predicted the capabilities are equal.
  • True detrimental (TN): PIC hashing accurately predicted the capabilities are completely different.
  • False optimistic (FP): PIC hashing incorrectly predicted the capabilities are equal, however they don’t seem to be.
  • False detrimental (FN): PIC hashing incorrectly predicted the capabilities are completely different, however they’re equal.

To make it a bit of simpler to interpret, we coloration the nice outcomes inexperienced and the unhealthy outcomes purple. We are able to characterize these in what is named a confusion matrix:

table1_04152024

For instance, here’s a confusion matrix from an experiment wherein I take advantage of PIC hashing to match openssl variations 1.1.1w and 1.1.1v when they’re each compiled in the identical method. These two variations of openssl are fairly related, so we might anticipate that PIC hashing would do properly as a result of lots of capabilities will probably be an identical however shifted to completely different addresses. And, certainly, it does:

table2_04152024

Metrics: Accuracy, Precision, and Recall

So, when does PIC hashing work properly and when does it not? To reply these questions, we will want a better solution to consider the standard of a confusion matrix as a single quantity. At first look, accuracy looks as if essentially the most pure metric, which tells us: What number of pairs did hashing predict accurately? This is the same as

For the above instance, PIC hashing achieved an accuracy of

You would possibly assume 99.9 p.c is fairly good, however when you look intently, there’s a refined downside: Most perform pairs are not equal. In keeping with the bottom reality, there are TP+FN = 344+1=345 equal perform pairs, and TN+FP=118,602+78=118,680 nonequivalent perform pairs. So, if we simply guessed that every one perform pairs have been nonequivalent, we might nonetheless be proper 118680/(118680+345)=99.9 p.c of the time. Since accuracy weights all perform pairs equally, it’s not essentially the most applicable metric.

As a substitute, we wish a metric that emphasizes optimistic outcomes, which on this case are equal perform pairs. This result’s in line with our aim in reverse engineering, as a result of figuring out that two capabilities are equal permits a reverse engineer to switch data from one executable to a different and save time.

Three metrics that focus extra on optimistic instances (i.e., equal capabilities) are precision, recall, and F1 rating:

  • Precision: Of the perform pairs hashing declared equal, what number of have been really equal? This is the same as TP/(TP+FP).
  • Recall: Of the equal perform pairs, what number of did hashing accurately declare as equal? This is the same as TP/(TP+FN).
  • F1 rating: This can be a single metric that displays each the precision and recall. Particularly, it’s the harmonic imply of the precision and recall, or (2∗Recall∗Precision)/(Recall+Precision). In comparison with the arithmetic imply, the harmonic imply is extra delicate to low values. Which means that if both the precision or recall is low, the F1 rating can even be low.

So, trying on the instance above, we are able to compute the precision, recall, and F1 rating. The precision is 344/(344+78) = 0.81, the recall is 344 / (344 + 1) = 0.997, and the F1 rating is 2∗0.81∗0.997/(0.81+0.997)=0.89. PIC hashing is ready to determine 81 p.c of equal perform pairs, and when it does declare a pair is equal it’s right 99.7 p.c of the time. This corresponds to an F1 rating of 0.89 out of 1.0, which is fairly good.

Now, you could be questioning how properly PIC hashing performs when there are extra substantial variations between executables. Let’s have a look at one other experiment. On this one, I examine an openssl executable compiled with GCC to 1 compiled with Clang. As a result of GCC and Clang generate meeting code in another way, we might anticipate there to be much more variations.

Here’s a confusion matrix from this experiment:

table3_04152024

On this instance, PIC hashing achieved a recall of 23/(23+301)=0.07, and a precision of 23/(23+31) = 0.43. So, PIC hashing can solely determine 7 p.c of equal perform pairs, however when it does declare a pair is equal it’s right 43 p.c of the time. This corresponds to an F1 rating of 0.12 out of 1.0, which is fairly unhealthy. Think about that you simply spent hours reverse engineering the 324 capabilities in one of many executables solely to seek out that PIC hashing was solely in a position to determine 23 of them within the different executable. So, you’ll be pressured to needlessly reverse engineer the opposite capabilities from scratch. Can we do higher?

The Nice Fuzzy Hashing Debate

Within the subsequent publish on this sequence, we’ll introduce a really completely different kind of hashing referred to as fuzzy hashing, and discover whether or not it may yield higher efficiency than PIC hashing alone.  As with PIC hashing, fuzzy hashing reads a sequence of bytes and produces a hash.  Not like PIC hashing, nevertheless, while you examine two fuzzy hashes, you’re given a similarity worth between 0.0 and 1.0, the place 0.0 means fully dissimilar, and 1.0 means fully related.  We are going to current the outcomes of a number of experiments that pit PIC hashing and a preferred fuzzy hashing algorithm in opposition to one another and study their strengths and weaknesses within the context of malware reverse-engineering.



Supply hyperlink

RELATED ARTICLES

LEAVE A REPLY

Please enter your comment!
Please enter your name here

- Advertisment -
Google search engine

Most Popular

Recent Comments