integrating intrusion alert information to aid forensic ... document… · proof 1 2 integrating...

17
UNCORRECTED PROOF 1 2 Integrating intrusion alert information to aid forensic explanation: An 3 analytical intrusion detection framework for distributive IDS 4 Bon K. Sy * 5 Queens College, Computer Science Department, 65-30 Kissena Blvd., Flushing, NY 11367, USA 6 Graduate Center/CUNY, Computer Science Department, 365 Fifth Ave., New York, NY 10016, USA 7 9 article info 10 Article history: 11 Received 8 December 2006 12 Received in revised form 3 March 2008 13 Accepted 23 January 2009 14 Available online xxxx 15 Keywords: 16 Probabilistic inference 17 Model discovery 18 Intrusion detection 19 Forensic analysis 20 21 abstract 22 The objective of this research is to show an analytical intrusion detection framework (AIDF) comprised of 23 (i) a probability model discovery approach, and (ii) a probabilistic inference mechanism for generating 24 the most probable forensic explanation based on not only just the observed intrusion detection alerts, 25 but also the unreported signature rules that are revealed in the probability model. The significance of 26 the proposed probabilistic inference is its ability to integrate alert information available from IDS sensors 27 distributed across subnets. We choose the open source Snort to illustrate its feasibility, and demonstrate 28 the inference process applied to the intrusion detection alerts produced by Snort. Through a preliminary 29 experimental study, we illustrate the applicability of AIDF for information integration and the realization 30 of (i) a distributive IDS environment comprised of multiple sensors, and (ii) a mechanism for selecting 31 and integrating the probabilistic inference results from multiple models for composing the most probable 32 forensic explanation. 33 Ó 2009 Published by Elsevier B.V. 34 35 36 1. Introduction 37 Current intrusion detection technology provides rich informa- 38 tion about an attack. However, the primary focus of intrusion 39 detection is on one-bit information about whether the security 40 has been compromised. Forensic analysis, on the other hand, at- 41 tempts to understand and explain what has happened to the secu- 42 rity environment, and how a security breach could happen [1–3]. 43 We present an analytical intrusion detection framework (AIDF) 44 with an underlying structure comprised of a probability model dis- 45 covery and inference mechanism. The purpose of AIDF is to bridge 46 intrusion detection with forensic analysis based on inferring and 47 integrating alert information from distributive IDS sensors; 48 whereas the outcomes of an inference are referred to as forensic 49 explanations. It is important to note that forensic explanations typ- 50 ically are not what one will be seeking as the final product of a 51 forensic analysis. Rather, forensic explanations are useful stepping 52 stones for an analyst to narrow the possible attack goals to model, 53 and provide intermediate snapshots on events useful for modeling 54 attacks in a forensic analysis. 55 The relationship between intrusion detection and forensic 56 explanation could be perceived as the following: an attack signa- 57 ture rule of a misuse based IDS such as Snort can be conceived as 58 a manifestation of an illegal activity, and different combinations 59 of n illegal activities can be represented by n-bit binary strings; 60 whereas each binary string pattern of the n-bit constitutes a foren- 61 sic explanation on the corresponding underlying illegal activities. 62 In other words, there is a high level linguistic description of a 63 forensic explanation, which has a semantic interpretation using 64 predict calculus that corresponds to the truthfulness of the n-bit 65 binary string pattern about the conjunctive occurrence of the sig- 66 nature rules. The utility of a forensic explanation is its potential 67 for offering a basis on security policy modification, and on improv- 68 ing intrusion prevention capability; e.g., an ‘‘offline incremental” 69 rule revision in the case of misuse based intrusion detection. 70 The significance of this research is a novel probabilistic based 71 approach for uncovering hidden information during the course of 72 intrusion detection useful for generating forensic explanations. 73 Specifically, we will illustrate how the novel approach is applied 74 to generating the most probable forensic explanation from 75 the intrusion alerts produced by distributive Snort IDS sensors 76 [4,5]. 77 In Section 2 a brief survey on forensics, and the relationship be- 78 tween forensic explanation and analysis are discussed. In Section 3 79 different types of intrusion detection technology and alarm alerts 80 are discussed. In Section 4 the details on signature rules for misuse 81 based intrusion detection and network monitoring are described. 82 In Section 5 an analytical intrusion detection framework (AIDF) is 83 introduced. We will discuss an algebraic constraint formulation 84 within the AIDF for deriving multiple models based on incomplete 85 observations, and an inference process for discovering hidden 1566-2535/$ - see front matter Ó 2009 Published by Elsevier B.V. doi:10.1016/j.inffus.2009.01.001 * Address: Queens College, Computer Science Department, 65-30 Kissena Blvd., Flushing, NY 11367, USA. E-mail address: [email protected] Information Fusion xxx (2009) xxx–xxx Contents lists available at ScienceDirect Information Fusion journal homepage: www.elsevier.com/locate/inffus INFFUS 338 No. of Pages 17, Model 5G 16 February 2009 Disk Used ARTICLE IN PRESS Please cite this article in press as: B.K. Sy, Integrating intrusion alert information to aid forensic explanation: An ..., Informat. Fusion (2009), doi:10.1016/j.inffus.2009.01.001

Upload: others

Post on 26-Jun-2020

5 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Integrating intrusion alert information to aid forensic ... document… · PROOF 1 2 Integrating intrusion alert information to aid forensic explanation: An 3 analytical intrusion

1

2

3

4

56

7

9

1011121314

151617181920

2 1

3536

37

38

39

40

41

42

43

44

45

46

47

48

49

50

51

52

53

54

55

56

57

Information Fusion xxx (2009) xxx–xxx

INFFUS 338 No. of Pages 17, Model 5G

16 February 2009 Disk UsedARTICLE IN PRESS

Contents lists available at ScienceDirect

Information Fusion

journal homepage: www.elsevier .com/ locate/ inf fus

OO

FIntegrating intrusion alert information to aid forensic explanation: Ananalytical intrusion detection framework for distributive IDS

Bon K. Sy *

Queens College, Computer Science Department, 65-30 Kissena Blvd., Flushing, NY 11367, USAGraduate Center/CUNY, Computer Science Department, 365 Fifth Ave., New York, NY 10016, USA

2223242526272829303132

a r t i c l e i n f o

Article history:Received 8 December 2006Received in revised form 3 March 2008Accepted 23 January 2009Available online xxxx

Keywords:Probabilistic inferenceModel discoveryIntrusion detectionForensic analysis

33

1566-2535/$ - see front matter � 2009 Published bydoi:10.1016/j.inffus.2009.01.001

* Address: Queens College, Computer Science DepaFlushing, NY 11367, USA.

E-mail address: [email protected]

Please cite this article in press as: B.K. Sy(2009), doi:10.1016/j.inffus.2009.01.001

ED

PRa b s t r a c t

The objective of this research is to show an analytical intrusion detection framework (AIDF) comprised of(i) a probability model discovery approach, and (ii) a probabilistic inference mechanism for generatingthe most probable forensic explanation based on not only just the observed intrusion detection alerts,but also the unreported signature rules that are revealed in the probability model. The significance ofthe proposed probabilistic inference is its ability to integrate alert information available from IDS sensorsdistributed across subnets. We choose the open source Snort to illustrate its feasibility, and demonstratethe inference process applied to the intrusion detection alerts produced by Snort. Through a preliminaryexperimental study, we illustrate the applicability of AIDF for information integration and the realizationof (i) a distributive IDS environment comprised of multiple sensors, and (ii) a mechanism for selectingand integrating the probabilistic inference results from multiple models for composing the most probableforensic explanation.

� 2009 Published by Elsevier B.V.

T

34

58

59

60

61

62

63

64

65

66

67

68

69

70

71

72

73

74

75

76

77

78

79

80

UN

CO

RR

EC1. Introduction

Current intrusion detection technology provides rich informa-tion about an attack. However, the primary focus of intrusiondetection is on one-bit information about whether the securityhas been compromised. Forensic analysis, on the other hand, at-tempts to understand and explain what has happened to the secu-rity environment, and how a security breach could happen [1–3].

We present an analytical intrusion detection framework (AIDF)with an underlying structure comprised of a probability model dis-covery and inference mechanism. The purpose of AIDF is to bridgeintrusion detection with forensic analysis based on inferring andintegrating alert information from distributive IDS sensors;whereas the outcomes of an inference are referred to as forensicexplanations. It is important to note that forensic explanations typ-ically are not what one will be seeking as the final product of aforensic analysis. Rather, forensic explanations are useful steppingstones for an analyst to narrow the possible attack goals to model,and provide intermediate snapshots on events useful for modelingattacks in a forensic analysis.

The relationship between intrusion detection and forensicexplanation could be perceived as the following: an attack signa-ture rule of a misuse based IDS such as Snort can be conceived as

81

82

83

84

85

Elsevier B.V.

rtment, 65-30 Kissena Blvd.,

, Integrating intrusion alert

a manifestation of an illegal activity, and different combinationsof n illegal activities can be represented by n-bit binary strings;whereas each binary string pattern of the n-bit constitutes a foren-sic explanation on the corresponding underlying illegal activities.In other words, there is a high level linguistic description of aforensic explanation, which has a semantic interpretation usingpredict calculus that corresponds to the truthfulness of the n-bitbinary string pattern about the conjunctive occurrence of the sig-nature rules. The utility of a forensic explanation is its potentialfor offering a basis on security policy modification, and on improv-ing intrusion prevention capability; e.g., an ‘‘offline incremental”rule revision in the case of misuse based intrusion detection.

The significance of this research is a novel probabilistic basedapproach for uncovering hidden information during the course ofintrusion detection useful for generating forensic explanations.Specifically, we will illustrate how the novel approach is appliedto generating the most probable forensic explanation fromthe intrusion alerts produced by distributive Snort IDS sensors[4,5].

In Section 2 a brief survey on forensics, and the relationship be-tween forensic explanation and analysis are discussed. In Section 3different types of intrusion detection technology and alarm alertsare discussed. In Section 4 the details on signature rules for misusebased intrusion detection and network monitoring are described.In Section 5 an analytical intrusion detection framework (AIDF) isintroduced. We will discuss an algebraic constraint formulationwithin the AIDF for deriving multiple models based on incompleteobservations, and an inference process for discovering hidden

information to aid forensic explanation: An ..., Informat. Fusion

Page 2: Integrating intrusion alert information to aid forensic ... document… · PROOF 1 2 Integrating intrusion alert information to aid forensic explanation: An 3 analytical intrusion

T

86

87

88

89

90

91

92

93

94

95

96

97

98

99

100

101

102

103

104

105

106

107

108

109

110

111

112

113

114

115

116

117

118

119

120

121

122

123

124

125

126

127

128

129

130

131

132

133

134

135

136

137

138

139

140

141

142

143

144

145

146

147

148

149

150

151

152

153

154

155

156

157

158

159

160

161

162

163

164

165

166

167

168

169

170

171

172

173

174

175

176

177

178

179

180

181

182

183

184

185

186

187

188

189

190

191

192

193

194

195

196

197

198

199

200

201

202

203

204

205

206

207

208

209

210

211

212

213

214

215

2 B.K. Sy / Information Fusion xxx (2009) xxx–xxx

INFFUS 338 No. of Pages 17, Model 5G

16 February 2009 Disk UsedARTICLE IN PRESS

UN

CO

RR

EC

inhibiting negative for a forensic analysis. In Section 6 the theoret-ical foundation behind the probability model discovery for theAIDF is detailed. In Section 7 the scenario of applying AIDF to thereal world situation for integrating intrusion alert informationfrom multiple sensors are discussed. The scalability issue of theproposed AIDF is presented in Section 8. An illustration on distrib-utive coverage to address the scalability issue is presented in Sec-tion 9. In Section 10 an experimental study and its evaluation arepresented, followed by the conclusion and future work in Section11.

2. Brief survey on forensics

Forensic analysis [6] is commonly perceived as ‘‘the process ofunderstanding, re-creating, and analyzing arbitrary events thathave previously occurred.” As of this writing, the field of computerforensics is still considered largely ad hoc [6]. Ideally, forensic anal-ysis will explain: (1) What has happened? (2) When did it happen?(3) How did it happen? (4) Who caused it? Furthermore, it wouldbe desirable if the forensic explanation is based on a systematicextraction and analysis of evidence that is admissible in a legalproceeding.

Unfortunately, as pointed out by Sommer [7], there is a gap be-tween the purposes of IDSs of various types and the needs of thelegal system. The gap exists not only at a purposive and functionallevel, but also in philosophical approach as to what is sufficient orconstitutes a ‘‘proof” of an intrusion offense. Therefore, in a legalproceeding, IDS alerts or alert correlation in an ESM (EnterpriseSecurity Management) system may contribute merely one step to-wards proving guilt of a criminal offense or liability in a civilmatter.

To the best of our knowledge, there is no one single intrusiondetection/prevention solution that can claim its alert evidencebeing sufficient and reliable, nor can it always satisfy the stringentrequirement of the chain of custody in order for it to be admissiblein a legal proceeding. This issue is further complicated by the inter-national jurisdiction when the source of an intrusion violation andthe targeted victim are from different countries. For example,forensic evidence is legally admissible in the US only if (1) it is per-missible (via lawful search/seizure), (2) preserved (following thechain of custody on how the evidence is collected, possessed, con-trolled, and made accountable), (3) trustworthiness (i.e., evidencecollected during normal business operation, not aftermath), and(4) relevant. The forensic aspect of this research is by no meanssufficient for legal admissibility. Rather, it is meant to help betterunderstand what and how could happen, thereby offers insightsinto preventing potential malicious attacks.

One active research area in this direction is the development ofmodels towards forensic analysis. For example, one idea is to de-velop the model for back tracking as exemplified by BackTracker[8]. BackTracker records system calls and processes in form of a‘‘dependency” graph. Such a ‘‘dependency” graph reviews thedependencies and traces of the sequence of the compromised sys-tems events. However, the criticism [9] on BackTracker is itsrequirement on an analyst to provide the system events to exam-ine. Yet the suspicious files and process IDs as related to potentiallycompromised system events may not be easily discovered if theanalysis takes place long after the intrusion.

Forensix [10] is another tool similar to BackTracker with thecapability of recording system calls. Although Forensix possessesan augmented feature to allow an analyst to make database queryon the activities of system calls, it is subjected to a similar con-straint; i.e., an analyst has to determine what specific files to inves-tigate based on her belief on whether such files may have beenmodified by an intruder’s code. In other words, the success of bothBackTracker and Forensix depends on the ability of an analyst to

Please cite this article in press as: B.K. Sy, Integrating intrusion alert(2009), doi:10.1016/j.inffus.2009.01.001

ED

PR

OO

F

pinpoint ‘‘root cause” in an ad hoc exploratory process of a forensicanalysis.

Peisert et al. [9] proposed an approach towards modeling foren-sic analysis – referred to as PB forensic approach hereafter in thispaper. PB approach models a generic attack in form of a three-layerdirected graph. The first layer is comprised of nodes representingevents that signify the beginning of an attack. The third layer iscomprised of nodes representing the end goals of an attack. Anend goal of an attack is defined as the success on achieving a par-ticular violation. The middle layer is comprised of many intercon-nected nodes; whereas each node in the middle layer represents anintermediate step of an attack and a unique traversal is a possiblerealization of an attack achieving the end goals of an intruder. Inother words, one unique traversal can be thought of as a revelationof ‘‘how could an attack happen.” The objective of a forensic anal-ysis is to identify the actual traversal path among many possibleones.

PB approach performs a goal-oriented attack modeling basedon the so-called requires/provides pre-conditions. Each directedpath encodes the temporal order of the system events occurredover time. A transition from one node to the next node signifiesthe ‘‘capabilities” in the requires/provides model. Their modelingapproach is based on two important assumptions. First, forensicanalysis is computationally tractable if one can carefully inte-grate the useful information at different abstraction layers (suchas low level system calls and high level alerts). In doing so, thesearch space of the path traversals in the middle layer can bekept manageable. Second, their model implicitly assumes abso-lute certainty, and sufficiency, on the ‘‘capability” attribute setfor a build-out of the nodes in the middle layer. As such, theirmodeling approach explicitly discards the notion of implication,which estimates the theoretical possibility of a hypothesizedoutcome.

In this research, we argue that extending the model for forensicanalysis to incorporate uncertainty can enhance its utility. Ourargument is based on three observations. First, a sophisticatedforensic model has to incorporate low level and high level events.The relationship between these events in different levels is oftenmany-to-many. One way to capture the potential ambiguity ofthe many-to-many relationship is through the concept of entropyin information theory, which is grounded on probability theory.Second, uncertainty measure can be useful in a forensic analysisto offer a more granular forensic explanation; e.g., without a studyon the level of consistency of certain time-varying statisticalbehavior, source/destination pair over time is not sufficient to deci-pher whether similar attacks are from the same or different parties.Third, a useful forensic model may render an exponentially largesearch space (in the middle layer). Proper probability assignmentcould improve the search efficiency by incorporating heuristicstrategy that may bring us closer to a realistic real-time forensicanalysis.

One natural way to generalize a forensic model for incorporat-ing uncertainty is to consider Markov modeling. The advantagesof extending and generalizing a forensic model to a kind of Markovmodel are (1) the existence of efficient search techniques (e.g., Ver-tibi-like search algorithm based on the joint probability of a tra-versal path) for a forensic analysis, and (2) the ability to reflectthrough uncertainty estimate the nature of many-to-many rela-tionship between the alert events that appear at different levelsof abstraction (e.g., packet stream characteristics such as protocoland payload contents on one extreme and the one-bit intrusionalert on the other extreme).

To realize the advantages just mentioned, we need to map theelements of a forensic model to that of a Markov model. An essen-tial step for the mapping is the task of deriving probability modelsfor the nodes characterizing the intermediate attack goals in the

information to aid forensic explanation: An ..., Informat. Fusion

Page 3: Integrating intrusion alert information to aid forensic ... document… · PROOF 1 2 Integrating intrusion alert information to aid forensic explanation: An 3 analytical intrusion

T

216

217

218

219

220

221

222

223

224

225

226

227

228

229

230

231

232

233

234

235

236

237

238

239

240

241

242

243

244

245

246

247

248

250250

251

252

253

254

255

256

257

258

259

260

261

262

263

264

265

266

267

268

269

270

271

272

273

274

275

276

277

278

279

280

281

282

283

284

285

286

287

288

289

290

291

292

293

294

295

296

297

298

299

300

301

302

303

304

305

306

307

308

309

310

311

312

313

314

315

316

317

318

319

320

321

322

323

324

325

326

327

328

329

B.K. Sy / Information Fusion xxx (2009) xxx–xxx 3

INFFUS 338 No. of Pages 17, Model 5G

16 February 2009 Disk UsedARTICLE IN PRESS

UN

CO

RR

EC

middle layer of a forensic model. This is necessary because eachnode in a Markov model encompasses a probability distribution(i.e., a probability model by itself). Time stamps of the events de-fine the temporal constraints on the order of the nodes representedby the transitional directed path in the model. Below is an examplerule set that will be used to illustrate the case in point and forensicexplanations:

2 BLEEDING-EDGE VIRUS HTTP Challenge/ResponseAuthentication

3 (http_inspect) OVERSIZE REQUEST-URI DIRECTORY5 BLEEDING-EDGE EXPLOIT Awstats Remote Code Execution

Attempt6 BLEEDING-EDGE EXPLOIT XML-RPC for PHP Remote Code

Injection47 BLEEDING-EDGE Exploit Suspected PHP Injection Attack48 BLEEDING-EDGE EXPLOIT WEB PHP remote file include

exploit attempt

Suppose rules 6 and 47 were reported in an intrusion alert, aforensic explanation with a scope covering these six rules has acorresponding 6-bit binary string pattern of 000110 that charac-terizes an intrusion event targeted at the vulnerability of PHP.As in another example, forensic explanation ‘‘web exploit vulner-ability” could have a semantic interpretation on a 6-bit binarystring corresponding to 100001. Any of such binary string withmatching access attributes (qualified by, for example, sourceand destination address pair, session/process ID, etc.) can be usedto represent a snapshot of an intermediate stage of an attack. Thesignature rules in the binary string can be mapped to low-levelsystem events, or used directly in modeling an attack during aforensic analysis session. Below shows an attack model basedon PB approach; where the directed arcs signify the order ofthe events occurred over time:

Time interval t1 – t2

R2

R3

R48

R6

R47

R5

Attack start

String pattern of forensic explanation with a scope of 6 rules: 110000 000001 000110 001000

Time interval t2 – t3

Time interval t3 – t4

Time interval t4 – t5

Attack end goal

Our research effort in this paper does not attempt to address the fullextent of a forensic analysis. Instead, our focus is on the snapshot ofeach time interval between ti and tj. Nevertheless, our effort con-tributes to extending a forensic model to incorporate uncertaintyby providing the followings:

1. Probability model discovery technique for deriving the proba-bility distribution of nodes in each time interval based on con-straints taking into the consideration of inhibiting negative – aconcept of alert alarms that will be presented in the nextsection.

2. Inferring missing information due to inhibiting negative thatmay reveal critical system events for modeling in a forensicanalysis.

3. Probabilistic inference technique for generating the forensicexplanations.

Please cite this article in press as: B.K. Sy, Integrating intrusion alert(2009), doi:10.1016/j.inffus.2009.01.001

ED

PR

OO

F

3. IDS and alert alarm types

Intrusion detection could be realized as host based or networkbased. This paper is focused on network based intrusion detection.Readers interested in the application layer intrusion detection arereferred to our previous work elsewhere [11]. Network basedintrusion detection could be broadly classified into two general ap-proaches; namely, misuse based or anomaly based [12,13,34].Anomaly based and misuse based intrusion detection each has itsown strength and weakness. As reported elsewhere [13], hybridapproach combining misuse based and anomaly based intrusiondetection has been proposed. One of the main thrusts on the hybridapproach is the potential on combining high-level alerts such asalert correlation or aggregation.

In order to fully realize the potential of hybrid approach forintrusion detection, it is desirable to have a rich language forabstracting and integrating high-level information from alert logsof multiple sensors. In particular, it is desirable to have a languagethat is based on multilevel abstractions and mechanisms forexpressing uncertainty in characterizing events, and for restructur-ing the representational space of events in ways that reduce thelikelihood of successful evasive attempts. As succinctly mentionedby Dolye et al. [14], the limitations of misuse and anomaly basedintrusion detection stem not from the methods themselves butfrom reliance on inadequate expressive languages for describingsignificant patterns.

The language issue just mentioned is a factor that affects thesuccess on event characterization and recognition during intrusiondetection. For example, Snort does not provide a convenient way toencode in a signature rule that can express an event such as ‘‘theprobability of a DoS (denial of service) attack in a subnet withtwo or more mission critical servers increases by more than 10%per hour over a period of 6 h.” Since the language such as theone in Snort has an inherent limitation on the expressivenessand semantics to describe temporally related events characterizedby uncertainty, the extent of possible success on integrating alertinformation to aid forensic analysis would have to be consideredwithin this scope. Obviously this research could not overcomethe limitation just mentioned without focusing on the develop-ment of a language for describing and characterizing intrusionevents. Nonetheless, this research attempts to explore one aspectof the forensic analysis permitted by the language of Snort;namely, event characterization that is based on linguistic sen-tences that are composed of the normal conjunction and disjunc-tion of the occurrence of activities captured in the signaturerules. As we have noted, the need for a possible cooperation be-tween IDS and forensic systems is timely in light of the industrialefforts in the development of forensic tools such as InfiniStream(by Network Associates) [15] and NetIntercept (by SandstormEnterprise) [16], as well as academic research effort on the devel-opment of ForNet [17].

Since Snort relies on signature rules for intrusion detection, theimplementation of its search engine must consider the possibilityof DoS attack due to an exploit on the nature of the multi-rule pat-tern matching [18]. In other words, one may exploit the nature ofmulti-rule pattern matching by deliberately sending network traf-fic that will trigger a large number of rules, and as such, overflowsthe processor and consumes the memory of an inspection engine.Current anti-DoS strategy [19] to prevent this potential exploit isto perform a set partition on the rules. The basic idea behind theset partition on the rules is to group the rules into categories suchas protocol field rules, generic content rules, packet anomaly rules,and IP rules. The inspection engine will first decide within whichcategory of the rule set to perform a search. It will then conducta search for matching rules against network traffic. Afterwards, it

information to aid forensic explanation: An ..., Informat. Fusion

Page 4: Integrating intrusion alert information to aid forensic ... document… · PROOF 1 2 Integrating intrusion alert information to aid forensic explanation: An 3 analytical intrusion

T

330

331

332

333

334

335

336

337

338

339

340

341

342

343

344

345

346

347

348

349

350

351

352

353

354

355

356

357

358

359

360

361

362

363

364

365

366

367

368

369

370

371

372

373

374

375

376

377

378

379

380

381

382

383

384

385

386

387

388

389

390

391

392

393

394

395

396

397

398

399

400

401

402

403

404

405

406

407

408

409

410

411

412

413

414

415

416

417

418

419

420

421

422

423

424

425

426

427

428

429

430

431

432

433

434

435

436

437

438

439

440

441

442

443

444

445

446

447

448

449

450

451

452

453

454

455

456

457

458

459

460

4 B.K. Sy / Information Fusion xxx (2009) xxx–xxx

INFFUS 338 No. of Pages 17, Model 5G

16 February 2009 Disk UsedARTICLE IN PRESS

UN

CO

RR

EC

will store the matching rules up to the limit of a queue with somepre-defined size, and report a small subset of the matching rulesstored in the queue. For example, the default configuration forSnort implementation has a queue size of eight; i.e., it stores upto eight matching rules in a queue even if there may be more thaneight matching rules. The default configuration also has a limit ofreporting up to three matching rules stored in the queue; in whichthe three rules may be chosen based on some pre-defined prioritycriteria, or in a probabilistic manner to prevent an attack thatanticipates the behavior of the search engine. In other words, ifthere are more than three rules in the queue, an end-user will onlybe able to observe up to three matching rules in the queue. As such,there are three different kinds of negative behavior of misuse basedintrusion detection:

1. True negative – When there is no attack and the misuse basedintrusion detection system reports no attack alert.

2. False negative – When there is an attack and the misuse basedintrusion detection system fails to detect and fails to generatean alert. This occurs when there is no signature rule to matchthe attack patterns, or the encoded signature rules fail due toeither the lack of scope coverage on the new attack or practicalprocessor limitation.

3. Inhibiting negative – When there is an attack and the misusebased intrusion detects, but fails, to report the attack. Thisoccurs when the set of encoded signature rules indeed detectsthe attack, but the inspection engine does not report all thematching rules because of the anti-DoS strategy. Referring tothe previous discussion, anti-DoS strategy imposes capturingand reporting limit. Consequently, when the reporting limit isreached, the matching rules that are reported inhibit othermatching rules to be reported.

While false negative is a concept that has been widely used forpossible evaluation of an intrusion detection approach, we alsorecognize its limitation in terms of its applicability to only simula-tion test. In the real world scenario, unless one could detect devi-ation of the behavior of the computing/networking environmentand could link the deviation to the external disturbance, false neg-ative will be very difficult, if not impossible, to detect. This is par-ticularly so in an environment where there is no additionalsource(s) of IDS data for cross reference. On the other hand, falsepositive is a concept that has been widely used for characterizingfalse alarms; i.e., intrusion detection system generates an alertwhen there is no attack. False alarms could be particularly prevail-ing and subject to ‘‘base rate fallacy” [20] if the signatures rules arenot properly tuned to the objective of the security policy to beenforced.

In this research, our focus is on inhibiting negative and some as-pects of false negative. The premise is that the network traffic pat-terns encoded in the signature rules could reveal information, ifcombined appropriately, to provide useful forensic explanations.Thus the objective is to uncover through a discovery process thehidden matching rules that are not observable, and to develop amodel through which probabilistic inference could be carried outto derive the most probable forensic explanation; whereas themodel incorporates the observable matching rules and the hiddenones that are uncovered.

One may argue that generating forensic explanation throughinhibiting negative is not necessarily viable. It is because, as tech-nology progresses, one may afford to set a large queue size and toincrease the number of matching rules to be reported while pre-serving anti-DoS strategy. In this case, the proposed researchmay not be valuable. While this may be acceptable from the prac-tical point of view at the present moment, this ‘‘quick fix” solutiononly works when one accepts the queue size to grow linearly with

Please cite this article in press as: B.K. Sy, Integrating intrusion alert(2009), doi:10.1016/j.inffus.2009.01.001

ED

PR

OO

F

respect to the number of matching rules. Irrespective to whetherthe size of the rule set follows Moore’s Law, it is pointed out else-where [21] that the simple technique of linearly searching throughthe set of rules is becoming increasingly infeasible.

Furthermore, one may wonder whether any existing IDS otherthan Snort may also be subject to inhibiting negative. As we havenoted, there are other IDS appliances based on Snort implementa-tion; e.g., NinjaBox Z-series system by Endace [22] and SourcefireIS3000 V4.0.2 [23]. While it is presented as an advantageous fea-ture, the technical manual of Sourcefire IS3000 system has alsopointed out the importance of determining the type of back-ground traffic and the maximum load that a sensor can sustain.This is emphasized because the system will begin to drop pack-ets/miss alerts when the maximum load is exceeded – suggestingthe potential issues caused by inhibiting negative. In other words,the issue is not whether there is a board array of IDS systems sub-ject to inhibiting negative. But rather the significance of the prob-lem is whether industry-standard IDSs are subject to inhibitingnegative.

Model-based probabilistic inference is proposed to addressinhibiting negative. However, the motivation and utility of proba-bilistic inference is to explore more than just the intrusion detec-tion methodology that decouples the dependency of itsperformance from the size of the rule set database. Rather, it hasadditional utilities from the real time perspective in regard to itsability to integrate information from the collaborative IDS sensorsthat realize a scaleable distributive IDS environment. This will be-come clearer after the analytical detection framework is intro-duced in Sections 5 and 7. Furthermore, the proposed techniquecould offer insights into evasive attempts of malicious attacks.We will show two such examples using Fragroute and web basedftp exploit.

Fragroute is a tool for manipulating packets of data so thatmalicious attacks could be camouflaged to bypass an intrusiondetection system. This tool has gained attentions few years agowhen it was released. The basic idea is to exploit the fact thatIDS does not perform data integrity check as stringent as a serversoftware. An example ‘‘insertion” attacks is to send a command toa server that could be disguised by adding extraneous, illegitimatedata.

Yet another example of an intrusion bypassing IDS is a simpleweb exploit such as http://ftp.sloppery_setup.com/pub/../../ In asloppy ftp site accessible via the http, if directory access protec-tion is not properly configured at the OS and server setup, ‘‘./pub/../../” allows one to traverse up to a directory level that oftenis not supposed to be shared with the public. Many such similarpatterns may exist for an exploit. It is a challenge for IDS to deter-mine the nature of the ‘‘unusual” activities and to catch this kindof patterns via simply an explicit encoding of the ‘‘patterns” assignatures.

We argue the proposed probabilistic inference for discoveringinhibiting negative could be easily extended to offer a possible ave-nue to address the above issues related to (1) assessing the ambi-guity of the ‘‘unusual activities” through likelihood assessment ofthe event from the discovered probability model(s), and (2) discov-ering these activities of which there may not even be any explicitsignature rule.

Using the ‘‘sloppy ftp site” example just discussed, one couldnote that the possible success on bypassing IDS is an exploit onthe lack of explicit signature rule encoding in layer 3. Yet theweb log at layer 7 would reveal abnormal directory listing and tra-versal. When such additional information is made available for thisapproach, one could construct a model that incorporates informa-tion made available from different sources. With such a model,probabilistic inference would become possible to help constructconjectures on possible evasive attempts of malicious attacks.

information to aid forensic explanation: An ..., Informat. Fusion

Page 5: Integrating intrusion alert information to aid forensic ... document… · PROOF 1 2 Integrating intrusion alert information to aid forensic explanation: An 3 analytical intrusion

461

462

463

464

466467

468

469

470

471

472

473

474475

476477

478

479

480

481

482

483

484

485

486

487488

489490

491492

493494

495

496

497

498

499

500

501

502

503

504

505

506

507

508

509

510

511

512

513

514

515

516

517

518

519

520

521

522

523

524

525

526

527

528

529

530

531

532

533

534

535

536

537

538

539

540

541

542

543

544

545

B.K. Sy / Information Fusion xxx (2009) xxx–xxx 5

INFFUS 338 No. of Pages 17, Model 5G

16 February 2009 Disk UsedARTICLE IN PRESS

4. Some details on snort signature rules

In this section we use Snort as a basis to explain the mechanismbehind the signature rules used for intrusion detection. Three sim-plified sample Snort signature rules are shown below:

Signature rule R1:

Please cite this article in press as: B.K. Sy, Integrating intrusion(2009), doi:10.1016/j.inffus.2009.01.001

546

alert tcp any any -> $HOME_NET 22 (msg: ‘‘SSH Successfuluser

n

547

548

connection”; dsize: 52; flags: AP; threshold: type both, trackby_src, count 3,

n

549

seconds 60; classtype: successful-user; sid: 2001637; rev:3;) 550 Signature rule R2: 551

552

alert tcp $EXTERNAL_NET any -> $HOME_NET 22 (msg:‘‘Potential SSH

n

553

554

Scan”; flags: S; flowbits: set, ssh.brute.attempt; threshold:type threshold,

n

555

556

track by_src, count 5, seconds 120; classtype: suspicious-login; sid: 2001219; rev:12;)

n

557

Signature rule R3: 558 alert tcp $EXTERNAL_NET 23 -> $HOME_NET any n 559 (msg: ‘‘TELNET client ENV OPT escape overflow attempt”; n 560 flow:to_client,established; content: ‘‘|FF FA|0|01|”; n 561

562

rawbytes; pcre:”/(nx02([ nx01 nx02 nx03]| nxFF nxFF)){100,}/RBsm”;

n

563

content: ‘‘|FF F0|”; distance:0; rawbytes;reference:bugtraq,12918;

n

564

reference:cve,2005-0469; classtype:attempted-user;sid:3537; rev:2;)

T

565

566

567

568

569

570

571

572

573

574

575

576

577

578

579

580

581

582

583

584

585

586

588589590

591592593

594595596

597598599

600601602

603604605

606

607

608

609

610

611

UN

CO

RR

EC

Briefly, rule R1 shows a user behavior access rule that belongs to theclass ‘‘successful-user.” This rule encodes a network traffic patternthat is connection-oriented using TCP with a packet payload size52, and the flow of the traffic is from any IP using any port in theInternet to a host in the local area network specified in $HOME_NETat port 22. In addition, this rule is triggered to send an alert only ifthe payload packet size is 52, and the TCP flag bits ‘‘ACK” and ‘‘PSH”are checked. Furthermore, the number of login attempts from thesource IP must be three times (or less) within the 60 s interval.

Rule R2, on the other hand, is an attack signature rule and be-longs to the class ‘‘suspicious-login.” As could be noted, R2 is mutu-ally exclusive to rule R1, while rule R3 can occur as a standalone, orin combination with R1 or R2.

When a Snort signature rule is triggered to send an alert, the log(using, for example, CSV format) will contain the informationabout <TimeStamp: date/hh24:min:sec> <Protocol> <Source IP ad-dress> <Source port> <Destination IP address> <Destination port><Signature rule>. Below is an example:

Dec-12-2005/03:43:01 TCP 139.103.65.163 54708149.4.211.170.236 22 R1

As the size of the signature rule set grows inevitably over timebecause of the discovery of the new attacks, one may incline tothink that the effectiveness of an IDS will remain so long as theprocessing power could keep up proportionally – thus providinga ground for re-examining the significance of the proposed AIDF.Unfortunately, this would be an over simplification on the relation-ship between the processing power and the size of signature ruleset. We will use a configurable feature of Snort to illustrate thispoint.

Snort could be configured to enable or disable different kinds ofpreprocessors such as stream4/5 and http. While Snort treats IP/TCP/UDP traffic in the packet level, Snort could be configured (viasnort.conf) to enable stream preprocessor to generate evasionalerts and/or to reassemble packets into larger ‘‘psuedopackets”

alert

ED

PR

OO

F

for an inspection. Similarly, Snort could be configured to invokehttp preprocessor to perform deep packet content inspection basedon various flow depth at predefined ports. For the sake of illustra-tion, let us consider the following signature rule in a rule setdatabase:

alert tcp any any -> any any (msg: ‘‘Test rule”; content:”someT-estString”; nocase; =classtype: attempted-user; sid: 2400000; rev 1;)

When the http preprocessor is enabled with default flow depth1460, an outbound URL request ‘‘http://www.google.com/someT-estString” will trigger the rule just shown above. However, if ‘‘som-eTestString” appears in the http response content section but notin the URL, the rule just shown would not necessary be triggered.It is because the maximum value of the flow depth (i.e., 1460) doesnot guarantee an inspection on the entire server response – unlessthe flow depth is set to assume a value of 0. However, when theflow depth parameter is set to 0, Snort could hang or crash depend-ing upon the type and the amount of traffic being inspected. In or-der for Snort to operate with sufficient stability, it is common to setthe flow depth to be 1460, even if it means to miss reporting a rulethat should have been triggered in an intrusion detection session.In other words, a database with even just one single rule could leadto inhibiting or false negative irrespective to the processing powerof an inspection engine.

5. Analytical intrusion detection framework

In our analytical intrusion detection framework (AIDF) that sup-ports forensic inference and explanation generation, the basis is abinary-valued table indicating any match between network datapacket and alert patterns encoded in a signature rule. As shownin the previous section, whenever there is one or more rule(s)matching the network traffic, Snort will report the network trafficinformation as well as the rule(s) being triggered. Below is a bin-ary-valued representation of an example Snort log file:

In Table 1, 0 indicates no match and 1 indicates a match of thecorresponding rule. Referring to the previous section, R1 and R2would not co-occur simultaneously. It could be formulated proba-bilistically as Pr(R1:1|R2:1) = 0. If the configuration is set to have alimit of reporting only one matching rule from the queue, then onewould not be able to observe the co-occurrence of R1 and R3. Thisis because only one of them will be reported even if there is amatch for both rules and the rules are stored in the queue. In otherwords, the observation of any row entry with a 1 does not neces-sarily reveal the true underlying phenomenon. The only situationthat the observation reveals the true underlying phenomenonwithout the ‘‘inhibiting factor” is a row with all zeros.

To further illustrate this, let us assume simple frequency countis used for estimating the occurrence likelihood of the data in Table1, and the following is obtained:

Pr(R1:1|R2:1) = 0

information to aid forensic

()

explan

PR3PrðR1 : 1;R2 : 1;R3Þ ¼ 0P

Pr(R1:1) P 150/30000

() R2,R3Pr(R1:1, R2, R3) P 0.005P Pr(R2:1) P 60/30000 () R1,R3 Pr(R1, R2:1, R3) P 0.002P Pr(R3:1) P 90/30000 () R1,R2 Pr(R1, R2, R3:1) P 0.003 Pr(R1:0 R2:0 R3:0) = 0.99P () Pr(R1:0, R2:0, R3:0) = 0.99

R1,R2,R3 Pr(R1, R2, R3) = 1

Table 2 shows six possible probability models consistent with theformulation just shown. These models are derived following theprobability model discovery technique discussed elsewhere [24],which will be detailed in the next section. Suppose an alert due torule 2 (SSH scan) is reported by the IDS, one may be interested in

ation: An ..., Informat. Fusion

Page 6: Integrating intrusion alert information to aid forensic ... document… · PROOF 1 2 Integrating intrusion alert information to aid forensic explanation: An 3 analytical intrusion

T

612

613

614

615

616

617

618

619

620

621

622

623

624

625

626

627

628

629

630

631

632

633

634

635

636

638638

639

640

641

642

643

644

645

646

647

648

650650

651

652

653

654

655

656

657

658

660660

661

663663

664

665

666

667

668

669

670

671

672

673

674676

677

678

679

680

681

682

683

684

685

686

Table 1Snort log in form of binary-valued table.

Timestamp R1 R2 R3

2005-12-12 03:43:01 0 1 0. . . . . . . . . . . .

2005-12-12 03:44:03 0 0 02005-12-12 03:44:04 1 0 02005-12-12 03:44:05 0 0 1

6 B.K. Sy / Information Fusion xxx (2009) xxx–xxx

INFFUS 338 No. of Pages 17, Model 5G

16 February 2009 Disk UsedARTICLE IN PRESS

NC

OR

REC

whether ESC overflow attempt (rule 3) also exists. In other words,one may be interested in ArgMaxR3[Pr(R3|R2:1)]. Let us assumeafter the models are derived, an additional piece of information isavailable: ‘‘ESC overflow attempt (R3) as a standalone activity with-out a successful SSH connection session as specified in R1 alwaysexceeds 0.4%.” Then only the last two models (marked with *) sat-isfy the additional condition Pr(R1:0|R3:1) P 0.004. In both cases,Pr(R3:1|R2:1) P Pr(R3:0|R2:1); i.e., the most probable explanationsuggests that whenever there is a SSH scan (R2), the activity ofESC overflow attempt may also occur. In other words, the mostprobable explanation based on the derived probability models al-lows us to deduce additional information about R3:1 (ESC overflowattempt) during a forensic inference session.

6. Probability model discovery for AIDF

In general, one can formulate a set of probability constraints formodel discovery based on the available information from IDS sen-sor(s). If only partial information is available because of, for example,inhibiting negative, the set of probability constraints defines an un-der-determined algebraic system. Each probability constraint can beexpressed as a linear combination of joint probability terms;whereas each joint probability term is treated as a non-negative un-known in the algebraic system. Probability model discovery is thenreduced to solving an under-determined algebraic system. Forexample, the probability constraint set shown right before Table 2in the previous section can be written in the matrix form below:

0 0 0 0 1 1 1 1 �1 0 00 0 1 1 0 0 1 1 0 �1 00 1 0 1 0 1 0 1 0 0 �10 0 0 0 0 0 1 1 0 0 01 0 0 0 0 0 0 0 0 0 01 1 1 1 1 1 1 1 0 0 0

266666664

377777775

pop1p2p3p4p5p6p7s0s1s2

266666666666666666664

377777777777777777775

¼

0:0050:0020:003

00:99

1

266666664

377777775

() A � x ¼ b; ð1Þ

where pi P 0, i = 0, . . . ,7, the binary representation of the index i inpi indicates the instantiation of R1, R2, R3, and sj (for j = 0, . . . ,2) arenon-negative slack variables. For example, p6 corresponds to theprobability term Pr(R1:1 R2:1 R3:0) because 610 = 1102.

U

Table 2Multiple probability model solutions.

Prob Solution model M: Variable instantiation for R1 R2 R3:

Pr(R1 R2 R3) 0 0 0 0 0 1 0 1 0

M1: Pr(R1 R2 R3) 0.99 0 0.002M2: Pr(R1 R2 R3) 0.99 0.001 0.002M3: Pr(R1 R2 R3) 0.99 0.002 0.002M4: Pr(R1 R2 R3) 0.99 0.003 0.002M5*: Pr(R1 R2 R3) 0.99 0.0025 0.000M6*: Pr(R1 R2 R3) 0.99 0.0017 0.000

Please cite this article in press as: B.K. Sy, Integrating intrusion alert(2009), doi:10.1016/j.inffus.2009.01.001

ED

PR

OO

F

The technique for solving the under-determined algebraic sys-tem is based on singular value decomposition (SVD), and whennecessary, the method developed by Kuenzi et al. [25] for obtainingan initial feasible solution. To illustrate the key idea behind the sin-gular value decomposition and the method by Kuenzi et al., thematrix form of an algebraic system is re-written as below:

A � x ¼ b() ½ai;j�i¼1;...;m;j¼1;...;n � x ¼ b: ð2Þ

Using the example in the previous section, m = 6 and n = 11. x is a11 � 1 vector and b is a 6 � 1 vector. The (i, j) entry of the matrixA, represented by ai,j, is �1, 0, or 1 in the example. The objectiveof the singular value decomposition is to decompose the matrix Ainto three matrices, U, D, and V, such that UDVT = A; whereas Uand V are orthogonal matrices and D is a diagonal matrix. A matrixM is orthogonal if MTM = MMT = I (identity matrix). Since U and V areorthogonal,

uk;j� �

k¼1;...;n;j¼1;...;m � ½uk0 ;j�k0¼1;...;m;j¼1;...;n ¼ I

()Xm

j¼1

uj;k0 � uj;k

" #k0¼1;...;n;k¼1;...;n

¼ I

()Xm

j¼1

uj;k0 � uj;k ¼0 when k0–k1 when k0 ¼ k

(

½vk0 ;j�k0¼1;...;n;j¼1;...;n � ½vk;j�k¼1;...;n;j¼1;...;n ¼ I

()Xn

j¼1

vk0 ;j � vk;j

" #k0¼1;...;n;k¼1;...;n

¼ I

()Xn

j¼1

vk0 ;j � vk;j ¼0 when k0–k

1 when k0 ¼ k

(

The mechanism behind the decomposition of A in SVD is to applyhouseholder transformation [26] to achieve QR decomposition[26] on A, and iteratively decompose the R matrix of the QR decom-position using QR iteration. The householder transformation as ap-plied to achieve QR decomposition guarantees the orthogonalproperty of the matrix Q. The goal of QR iteration is to iterativelydecompose the R matrix from triangular matrix to bi-diagonal ma-trix, and then further from bi-diagonal matrix to the diagonal ma-trix D as desired in the SVD. The following steps summarize theQR iteration for achieving singular value decomposition:

Step 0:Initialize A0 = A.

Step 1:Apply QR decomposition on A0; i.e., A0 = QUiRi.; where i initiallyequals 1 and represents the iteration index.

Step 2:Apply QR decomposition on the transpose of Ri (if it is notalready diagonal at i > 1); i.e., ðRiÞT ¼ QViR

0i.

Step 3:Determine whether R0Ti is diagonal. If R0Ti is diagonal, then theiteration is completed, otherwise let A0 ¼ R0Ti and go to step 1.

0 1 1 1 0 0 1 0 1 1 1 0 1 1 1

0.003 0.005 0 0 00.002 0.005 0 0 00.001 0.005 0 0 00 0.005 0 0 00.0025 0.005 0 0 00.0033 0.005 0 0 0

information to aid forensic explanation: An ..., Informat. Fusion

Page 7: Integrating intrusion alert information to aid forensic ... document… · PROOF 1 2 Integrating intrusion alert information to aid forensic explanation: An 3 analytical intrusion

687688

689

690

691

692

693

694

695

696

697

698

699

700

701

702

703

704

705

706

707

708

709

710

711

713713

714

715

716

717

718

719

720

721

722

723

724

725

726

727

728

729

730

731

732

733

734

735

B.K. Sy / Information Fusion xxx (2009) xxx–xxx 7

INFFUS 338 No. of Pages 17, Model 5G

16 February 2009 Disk UsedARTICLE IN PRESS

When R0Ti becomes a diagonal matrix at the completion of the kth QRiteration, A ¼ ð

QQ ÞR0Tð

QQT Þ ¼ ð

QQ ÞR0Tð

QQ ÞT . Since each

T

736

737

738

739

740

741

742

743

744

745

746

747

748

749

751

752

753

754

755

756

757

758

759

760

761

762

763

764

765

766

767

768

769

770

771

772

773

774

775

776

777

778

779

780

781

782

783

784

785

786

787

788

789

790

791

792

793

794

795

796

797

UN

CO

RR

EC

i Ui i i¼k Vi i Ui i i¼1 Vi

QUi and QVi is orthogonal, iQUi and kQVi are orthogonal. iQUi can then beused to define the U matrix of the SVD. i=kQVi can then be used to de-fine the V matrix of the SVD. And R0Ti is used to define the diagonal ma-trix D of the SVD. A complete detailed illustration of applyinghouseholder transformation and QR decomposition to realize the sin-gular value decomposition of A could be found online elsewhere [27].

Upon the completion of the singular value decomposition, onecan easily obtain the inverse of D by taking the reciprocal of thenon-zero diagonal entry in D while zero diagonal entry remainszero. Once U, V, and D�1 are known, solving A x ¼ b. is equivalentto solving UDVT x ¼ b. One can then easily derive x ¼ VD�1UT b.

Recall that the probability constraint set A is often an under-determined algebraic system. Many solutions may exist in themodel discovery process. We will now show an important propertyof SVD that permits efficient derivation of the multiple solutions.

Theorem 1. Let UDVT be the singular value decomposition of A. If theith diagonal entry of the D matrix is zero, then the correspondingcolumn vector Vi in the V matrix is a null vector with respect to A; i.e.,AVi = 0.

Proof. For A ¼Pm

k¼1ui;kdk;kvk;j

� �i¼1;...;m;j¼1;...;n, let the k0th diagonal

entry of D is zero; i.e., dk0k0 = 0. We need to prove AVk0 = 0; whereasV 0k ¼ ½vk0 ;i�i¼1;...;n is a column vector.

AVk0 ¼Xm

k¼1

ui;kdk;kvk;j

" #i¼1;...;m;j¼1;...;n

½vk0 ;i�i¼1;...;n

¼Xn

j¼1

Xm

k¼1

ui;kdk;kvk;j

!vk0 ;j

" #i¼1;...;m

¼Xn

j¼1

Xm

k¼1;k–k0ui;kdk;kvk;jvk0 ;j

24

35

i¼1;...;m

þXn

j¼1

ui;k0dk0 ;k0vk0 ;jvk0 ;j

" #i¼1;...;m

þXm

k¼1;k–k0ui;kdk;k

Xn

j¼1

vk0 ;jvk;j

!24

35

i¼1;...;m

þXn

j¼1

ui;k0dk0 ;k0v2k0 ;j

" #i¼1;...;m

By orthogonal definition stated earlier, the first term on the right-hand side above is a zero vector because

Pnj¼1vk0 ; j � vk;j ¼ 0. The sec-

ond term on the right-hand side above is also a zero vector becausedk0k0 is zero. Therefore, AVk0 is a zero vector. This completes theproof. h

By the concept of eigenvector, if A is an under-determined alge-braic system, some k0 diagonal entry/entries would be zero; i.e.,P

k0AVk0 ¼ 0() AðP

k0ck0Vk0 Þ ¼ 0 for some arbitrary constant ck0.Since Ax ¼ yA, y ¼ b; where y ¼ xþ

Pk0ck0mk0 . By changing the con-

stant ck0 in the linear combination, we obtain different values of y:The multiple probability model solutions can then be obtainedfrom those solutions of that are non-negative.

In applying the Singular Value Decomposition to discoveringmultiple probability models, all joint probability terms and theslack variables, if any, should be non-negative. Slack variables al-low the conversion of inequality constraints to equality con-straints. In the worst case scenario, we will need to produce aninitial probability model solution that satisfies the non-negativityrequirement if the initial solution x derived from the SVD failsthe non-negativity requirement. Once an initial solution satisfies

Please cite this article in press as: B.K. Sy, Integrating intrusion alert(2009), doi:10.1016/j.inffus.2009.01.001

ED

PR

OO

F

the non-negativity requirement, the process of generating multipleprobability model solutions based on the initial solution and thelinear combination of the null vectors extracted from the columnvectors in the matrix V of SVD can proceed as just described. In or-der to make sure that the probability model discovery process willnot fail when a set of proper probability constraints is presented,the method by Kuenzi et al. is used to generate an initial solutionwhen the worst case scenario occurs. The detail of the mechanismbehind the method by Kuenzi et al. will not be discussed in this pa-per because it has been extensively discussed elsewhere [25].Nonetheless, an example illustration is provided in Appendix 2.

For complete details on applying Singular Value Decompositionand the method by Kuenzi et al. for generating multiple models,readers are referred to Chapter 9 of [24] as well as [27]. Nonethe-less, the algorithmic steps for the model discovery are summarizedas below for the completeness of this paper:

Step 1:Construct the algebraic system of equations defined by theprobability constraint set in the form of Ax = b; i.e., eachconstraint, with proper slack variable introduced whennecessary, accounts for a row in matrix A.

Step 2:Apply the Singular Value Decomposition (SVD) algorithm onthe matrix A and obtain a feasible solution. If the solution ofthe SVD does not satisfy non-negativity requirement, applythe numerical method proposed by Kuenzi et al. to obtain afeasible solution that satisfies non-negativity requirement.

Step 3:Identify column vectors {Vk0|k0 = 1,2. . .} in the matrix V thatcorrespond to the zero entries in the diagonal matrix D of theSVD of A; where the SVD of A = UDVT, U and V are orthogonalmatrices.

Step 4:Obtain multiple alternative solutions y by constructing thelinear combination of the initial solution x with the Vk0; i.e., Ay= b where y = x + Rk0ck0Vk0 for some constants ck0. Retain thosesolutions satisfying non-negativity requirement.

The pivoting mechanism of Kuenzi et al. approach applied in step 2is quite common in linear programming (LP) solutions such as Sim-plex. However, our algorithm extends beyond solving linear optimi-zation problems. One of the key unique features of our algorithm isa ‘‘speculative fast” search based on SVD. This is not part of the LPsolutions. If in the worst case the SVD ‘‘speculative fast” searchcould not derive a solution satisfying the non-negativity con-straints, then both our and LP solutions would have to revert backto the selection of a pivot element based on incremental improve-ment of a modified cost function. However, LP is restricted tosearching only the vertices of the polytope (solution space), whichcannot reach any optimal solution that does not reside at some ver-tex of the polytope. On the other hand, the SVD approach uses thelinear combination of the null vector (step 4) to search throughthe inner space of the polytope using a linear path trajectory. There-fore, our approach overcomes the limitation of the traditional LPsolutions, and is capable of solving a non-linear optimization prob-lem where the best solution does not reside at a vertex.

The complexity of the SVD described above is in the order ofO(mnr); where m and n are the number of rows and columns in am � n matrix, and r is the rank of the matrix (proportional to thenumber of nonzero column entries). However, the number of un-known joint probability terms for model discovery grows exponen-tially with respect to the number of observed triggering rules.Therefore, discovering joint probability models using the SVD is

information to aid forensic explanation: An ..., Informat. Fusion

Page 8: Integrating intrusion alert information to aid forensic ... document… · PROOF 1 2 Integrating intrusion alert information to aid forensic explanation: An 3 analytical intrusion

T

798

799

800

801

802

803

804

805

806

807

808

809

810

811

812

813

814

815

816

817

818

819

820

821

822

824824

825

826

827

828

829

830

831

832

833

834

835

836

837

838

839

840

841

842

843

844

845

846

847

848

849

850

851

852

853

854

855

856

857

858

859

860

861

862

863

864

865

866

867

868

869

870

871

872

873

874

875

876

877

878

879

880

881

882

883

884

885

886

887

888

889

890

891

892

893

894

895

896

897

898

899

900

901

902

903

904

905

906

907908

909

910

911

912

8 B.K. Sy / Information Fusion xxx (2009) xxx–xxx

INFFUS 338 No. of Pages 17, Model 5G

16 February 2009 Disk UsedARTICLE IN PRESS

UN

CO

RR

EC

exponentially hard with respect to the number of joint probabilityunknowns. Although the nature of the problem is exponentiallyhard, the SVD based approach is well within the range for its appli-cation to the discovery of probability models of intrusion alerts. Aswe have experimented and reported elsewhere [24], the SVD, to-gether with the method by Kuenzi et al., have been successfully ap-plied to discover probability models for solving many algebraicproblems with an average size of 200–300 unknowns and a setof 12–24 (in)equality constraints. The most difficult problem beingsolved is one with 2187 unknowns and 10 constraints. When solv-ing an algebraic problem of this size, the model discovery processcan be completed almost instantly in a laptop with a 1.6 GHz CPUand 512 MB RAM using standard error tolerance 10�6 and conver-gence iteration limit 30 as suggested elsewhere [28,29]. Specifi-cally, the average time taken for solving the algebraic problem is14.71 s for problems with 16 unknowns or less, 17.99 s for thatwith 128 unknowns, 22.12 s for that with 256 unknowns, and86.32 s for that with 2187 unknowns. The average time just re-ported includes the time taken for a human user to interact withthe graphical user interface of the software, as well as the time ta-ken for a network connection and the authentication/authorizationprocess (for using our software). The plot below shows the rela-tionship among the log of the run-time (in seconds), the log ofthe number of unknowns, and the number of constraints, for differ-ent sizes of the algebraic problems:

In addition to relying on only the SVD decomposition to handle theexponential nature of the problem from the aspect of algorithmicefficiency and computing resources, opportunities also exist to re-duce the exponential problem to sub-exponential by consideringthe possible independency among the signature rules. Furthermore,we will explain in the latter section how the rule set partition in theAIDF could also improve the scalability of the model discovery bytaking advantage on the rule set partition to be distributed acrossmultiple IDS sites rather than at one aggregated IDS site. This is inaddition to using domain knowledge (if desired) on the possibleindependence and the mutually exclusiveness of the rules for deriv-ing set partition for the detection engine.

7. AIDF for integrating alert information from multiple sensors

The analytical intrusion detection framework is introduced inSection 5 for discovering probability model and for inferring hid-den alerts that are not reported due to inhibiting negative. How-ever, AIDF has a broader impact beyond just identifyinginhibiting negative. We could apply the probabilistic approach ofthe AIDF for integrating information from different sensor sources,not just Snort, in a distributive Network based IDS (NIDS)environment.

Typically, it is desirable for the placement of only minimal NIDSsensors for the maximal coverage of a network topology. Let us as-sume there is a network environment of three subnets SN1, SN2,and SN3. The placement of a single IDS sensor (with a database

Please cite this article in press as: B.K. Sy, Integrating intrusion alert(2009), doi:10.1016/j.inffus.2009.01.001

ED

PR

OO

F

of, say, 3000 rules) at a common network gateway/router may besufficient to monitor cross subnet traffic activities. However, anylocal unicast traffic activities based on, for example, MAC addressinformation stored in the ARP cache, could bypass the IDS sensor.On the other hand, host based IDS approach may not be effective– in terms of rule matching performance as well as the cost ofimplementation. It is because any host-to-host packet traversalin the same subnet will be examined at least twice. And if the pack-et traverses cross subnet, it may be examined once more if a NIDSsensor is placed at the layer 3 edge device (e.g., common gateway/router).

One alternative is to consider a distributive IDS environment. Ina distributive IDS environment, a small number of sensors could beplaced to collectively cover a local (subnet) environment. Each sen-sor is configured to handle a partial set of rules. Specifically, wewould like to leverage on the information gathered from individualsensors for reconstructing the activities that may otherwise re-quire a costly placement of extensive number of multiple IDS sen-sors. That is, we may opt to place n IDS sensors, as opposed to onesensor, in a subnet with m hosts; where n�m. For example, sup-pose m is 20 and n is 3, each IDS sensor is loaded with, say, 1200(out of the 3000) rules with 300 overlapping rules between a des-ignated pair of sensors. Any packet traversal broadcasting over thesubnet will be received by all three IDS sensors, and altogetherthey cover the space of all 3000 rules. In addition, any unicastpacket traversal will have 28% chance (assuming uniform distribu-tion) reaching at least one of the three sensors.

However, there is one scenario the distributive IDS strategy mayfail; i.e., a unicasting packet crafted for an attack targets at a victimsensor (referred to as the ignorant sensor) that has no matchingsignature rule(s). Yet the matching signature rule(s) exist(s) at an-other sensor (referred to as the partner sensor). A sensor is a ‘‘part-ner” of another sensor if overlapping rules exist between these twosensors. Our proposed approach provides a technique that couldaddress the drawback of the scenario just mentioned. Specifically,the attack just mentioned could be identified through a probabilis-tic inference using the model encapsulating the coupling/associa-tion statistical characteristics of the ignorant and partner sensorsas defined by the overlapping rules.

In general, the technique described in the AIDF provides a meth-od for integrating information from different IDS sensor sources,and could be applied to realize the IDS sensor collaboration not justacross local subnets, but wide area networks in which IDS sensorsare equipped with different signature rules. Deploying IDS sensorsequipped with different signature rules will be detailed later in the‘‘experimental study” section.

8. Scalability and distributive coverage of AIDF

Although probability model discovery process is just an exerciseon solving underdetermined algebraic system, the number of un-known joint probability terms grows exponentially with respectto the number of signature rules. Scalability is a complexity issuethat needs to be addressed for practical purpose.

To address the scalability issue, we introduce a distributive cov-erage framework. Let R be a set of n signature rules {R1, . . . ,Rn}; i.e.,|R| = n. Let R0 be a small subset of R (i.e., |R0| = m� n). With respectto R0, let GR0 = {G1R0, . . . ,GkR0} be k(>1) subsets of R0 satisfying the fol-lowing conditions:

1. [kj¼1GjR0 ¼ R0.

2. For each Rr R0, Rr�\iGj0R where each GjR0 GR0 ,r and GR0 ,r is a subsetof GR0 with |GR0 ,r| = s for some s > 1.

In a typical real world scenario for Snort, n is in the order of3000; i.e., n 3000. Suppose m = 8, k = 5, s = 3, the subset R0 consists

information to aid forensic explanation: An ..., Informat. Fusion

Page 9: Integrating intrusion alert information to aid forensic ... document… · PROOF 1 2 Integrating intrusion alert information to aid forensic explanation: An 3 analytical intrusion

913

914

915

916

917

918

920920

921923924

925926

927928

929930

931932

933934

935936

937938

939

940

941

942

943

944

945

946

947

948

949

950

951

952

953

954

955

956

957

958

959

960

961

962

963

964

965

966

967

968

969

970

971

972

973

974

975

976

977

978

979

980

981

Table 3Frequency count summary information.

R1 R2 R3 R4 R5 Count

0 1 0 1 0 10 0 1 0 1 3081 0 0 0 0 670 0 0 0 0 1,557,858

B.K. Sy / Information Fusion xxx (2009) xxx–xxx 9

INFFUS 338 No. of Pages 17, Model 5G

16 February 2009 Disk UsedARTICLE IN PRESS

of eight signatures out of the set of 3000 some rules. R0 typically isthe set of signature rules reported in intrusion alerts of a shortmoving time window. In the set GR0 there are five elements, andeach is a subset of R0. For each rule Rr in R0, it also appears as an ele-ment in three of the five subsets of R0 in GR0. Below shows an exam-ple of the formulation just described:

R0 ¼ R1; . . . ;R8; G0R ¼ G1R0 ; . . . ;G5R0 ;

G1R0 ¼ R1;R2;R3;R6;R8; G2R0 ¼ R2;R3;R4;R6;R8;G3R0 ¼ R3;R4;R5;R7;R8; G4R0 ¼ R1;R4;R5;R7;G5R0 ¼ R1;R2;R5;R6;R7;

R0 ,1

982 R1 G1R0 \ G4R0 \ G5R0 = {R1}

U

Table 4Probability models of different scope coverage.

Variable Instantiation 000

M(G1R0):Pr(R1 R2 R3) 0.999759M(G2R0):Pr(R1 R3 R4) 0.999759M(G3R0):Pr(R1 R3 R5) 0.999759M(G4R0):Pr(R1 R4 R5) 0.999759M(G5R0):Pr(R2 R3 R4) 0.999802M(G6R0):Pr(R3 R4 R5) 0.999802

Please cite this article in press as:(2009), doi:10.1016/j.inffus.2009.01

where G = {G1R0,G4R0,G5R0}0

983

R2 G1R0 \ G2R0 \ G5R0 = {R2,R6} where GR ,2 = {G1R0,G2R0,G5R0}0

984

R3 G1R0 \ G2R0 \ G3R0 = {R3,R8} where GR ,3 = {G1R0 ,G2R0 ,G3R0}0

985

R4 G2R0 \ G3R0 \ G4R0 = {R4} where GR ,4 = {G2R0,G3R0,G4R0} R5 G3R0 \ G4R0 \ G5R0 = {R5,R7} where GR0 ,5 = {G3R0 ,G4R0 ,G5R0}

0

R6 G1R0 \ G2R0 \ G5R0 = {R2,R6} where GR ,6 = {G1R0,G2R0,G5R0}

0

R7 G3R0 \ G4R0 \ G5R0 = {R5,R7} where GR ,7 = {G3R0,G4R0,G5R0}

0

987987 R8 G1R0 \ G2R0 \ G3R0 = {R3,R8} where GR ,8 = {G1R0,G2R0,G3R0}

T

988990

991

992

993994

995

996

997

998999

1000

10011002

1003

1004

1005

1006

1007

1008

1009

1010

1011

1012

1013

1014

1015

1016

1017

1018

NC

OR

REC

As shown in the previous section, the exponential complexity forthe probability model discovery depends on |GjR0|. Therefore, |GjR0|should be kept small for scalability reason. In the above example,|GjR0| = 5 for j = 1, . . . , 5 except 4, and |G4R0| = 4. The size of |GjR0|can remain small by increasing the size of |GR0| when the numberof rules increases. In addition, it is desirable to keep the size of |GR0 ,r|for different rules similar to each other so that the coverage for eachrule has minimal bias.

Recall an intrusion alert reported by Snort is in form of a ruletrigger. When a rule Rr is triggered, its corresponding cover set GR0 ,r

defines a number of GjR0 that each forms a basis for probabilitymodel discovery. In other words, probability model(s) can be de-rived for each GjR0 GR0 ,r, and probabilistic inference can be con-ducted on each GjR0 to derive the most probable explanationE[GjR0|Rr] based on ArgMax[Pr(GjR0|Rr)]. Within the distributive cov-erage framework, a forensic explanation is composed out of all themost probable explanations E[GjR0|Rr].

9. An illustration on distributive coverage

We illustrate the concept of distributive coverage frameworkjust discussed for forensic inference using one-week log data fromone of our local subnets. Specifically, the distributive coverageframework is applied to scale the complexity for discovering multi-

001 010 0

0 7E�07 06.42E�07 1.98E�04 04.30E�05 4.30E�05 11.98E�04 0 06.31E�07 1.97E�04 06.56E�07 0 0

B.K. Sy, Integrating intrusion alert.001

DPR

OO

F

ple models for one IDS sensor. Discovering multiple models formultiple IDS sensors will be illustrated in the next section. In thisillustration, a Snort sensor was installed in a local subnet192.168.0.0/24. A common configuration was used for the Snortinstallation; i.e., the queue size of Snort is eight, while the report-ing limit is two. The rule set of the experiment is the latest releaseprior to November 28, 2005 from the official Snort [30] and Blee-dingSnort [31] web sites. Altogether there are about 3000 ruleswith automatic daily update.

The testing period of the experiment is from November 28,2005 to December 5, 2005. During the testing period, five rulesshown below are triggered:

R1: Potential SSH ScanR2: SCAN NMAP – f – sSR3: Oversize Request – uridR4: SCAN NMAP – sSR5: Bare Byte Unicode.

The frequency occurrence of each rule is shown below with zeroreferring to no rule match while one referring to a rule match.

In Table 3, the traffic volume is the sum of all the counts on therightmost column. Referring to the distributive coverage frame-work formulation discussed in the previous section with m = 5,k = 6, s = 2, . . . ,5.

R0 ¼ R1; . . . ;R5; GR0 ¼ G1R0 ; . . . ;G6R0 ;

G1R0 ¼ R1;R2;R3; G2R0 ¼ R1;R3;R4; G3R0 ¼ R1;R3;R5;G4R0 ¼ R1;R4;R5; G5R0 ¼ R2;R3;R4; G6R0 ¼ R3;R4;R5;

R0 ,1

ER1 G1R0 \ G2R0 \ G3R0 \ G4R0

= {R1}

11 100 101

4.3E�05 1.98E�0 0

.12E�04 0 00 4.24E�0 06.56E�07 1.96E�

information to aid forensic expla

where G = {G1R0,G2R0,G3R0,G4R0}

R2 G1R0 \ G5R0 = {R2}

where GR0 ,2 = {G1R0,G5R0} R3 G1R0 \ G2R0 \ G3R0 \ G5R0

\ G6R0 = {R3}

where GR0 ,3 = {G1R0 ,G2R0 ,G3R0 ,G5R0 ,G6R0}

R4 G2R0 \ G4R0 \ G5R0 \ G6R0 = {R4}

whereGR0 ,4 = {G2R0,G4R0,G5R0,G6R0}

0

R5 G3R0 \ G4R0 \ G6R = {R5} where GR ,5 = {G3R0,G4R0,G6R0}

Under the condition of the reporting limit = 2, the constraint set forthe model M(G1R0) covering the scope (R1, R2, R3) becomes:

Pr(R1:1) P 67/1558234 = 0.0000429974Pr(R2:1) P 1/(1 + 308 + 67 + 1557858) = 0.0000006418Pr(R3:1) P 308/(1 + 308 + 67 + 1557858) = 0.0001976(67 + 309)/(1 + 308 + 67 + 1557858) = 376/1558234 =0.0002413 P Pr(R1:1)(1 + 308)/1558234 = 0.0001983 P Pr(R2:1)(1 + 308)/1558234 = 0.0001983 P Pr(R3:1)Pr(R1:1 R2:0 R3:0) = 67/(1 + 308 + 67 + 1557858) = 0.000043Pr(R1:0 R2:0 R3:0) = 1557858/(1 + 308 + 67 + 1557858) = 0.9997587

The upper bound for Pr(R1: 1) is arrived by observing 67 timesoccurrence of R1:1, plus the number of potential inhibiting nega-

110 111

04 0 04.30E�05 00 4.30E�05

05 0 6.42E�076.31E�07 1.07E�08

04 0 6.42E�07

nation: An ..., Informat. Fusion

Page 10: Integrating intrusion alert information to aid forensic ... document… · PROOF 1 2 Integrating intrusion alert information to aid forensic explanation: An 3 analytical intrusion

T

1019

1020

1021

1022

1023

1024

1025

1026

1027

1028

1029

1030

1031

1032

1033

1034

1035

1036

1037

1038

1039

1040

1041

1042

1043

1044

1045

1046

1047

1048

1049

1050

1051

1052

1053

1054

1055

1056

1057

1058

1059

1060

1061

1062

1063

1064

1065

1066

1067

1068

1069

1070

1071

1072

1073

1074

1075

1076

1077

1078

1079

1080

1081

1082

1083

1084

1085

1086

1087

1088

1089

1090

1091

1092

1093

1094

1095

Table 5Best explanation results from probabilistic inference.

Model Se Best Explanation satisfying(1) and(2)

M(G1R0) R1:1 R3:1 ArgMax[Pr(R1,2,R3|R1 = 1,R3 = 1)] = R2:0M(G2R0) ArgMax[Pr(R1,R3,R4|R1 = 1,R3 = 1)] = R4:0M(G3R0) ArgMax[Pr(R1,R3,R5|R1 = 1,R3 = 1)] = R5:1M(G4R0) ArgMax[Pr(R1,R4,R5|R1 = 1)] = R4:0 R5:1M(G5R0) ArgMax[Pr(R2,R3,R4|R3 = 1)] = R2:0 R4:0M(G6R0) ArgMax[Pr(R3,R4,R5|R3 = 1)] = R4:0 R5:1

INTERNET

Router/ firewall

149.4.211.0

Gateway/router Alert log

IDS sensor RL=2, QS=8

Signature rules

Computer G Computer H

67.93.250.0

10 B.K. Sy / Information Fusion xxx (2009) xxx–xxx

INFFUS 338 No. of Pages 17, Model 5G

16 February 2009 Disk UsedARTICLE IN PRESS

OR

REC

tive – hereafter referred to as inhibiting zero – on R1 due to thepatterns of at least two ‘‘1”s in the log data. These patterns are(R1:0 R2:1 R3:0 R4:1 R5:0) occurring once and (R1:0 R2:0 R3:1R4:0 R5:1) occurring 308 times. Similarly, the upper bound forPr(R2: 1) is due to the one-time occurrence of R2:1, plus the num-ber of potential inhibiting zero resulted from the pattern (R1:0R2:0 R3:1 R4:0 R5:1) occurring 308 times. And the upper boundfor Pr(R3: 1) is due to 308 times occurrence of R3:1, plus the num-ber of potential inhibiting zero resulted from the pattern (R1:0R2:1 R3:0 R4:1 R5:0) occurring once. Note that the pattern (R1:1R2:0 R3:0 R4:0 R5:0) cannot introduce inhibiting zero when thereporting limit is two. In general, the number of potential inhibit-ing zero for a rule variable Ri is the sum of the number of occur-rences of all patterns that has Ri being 0, and there are at least h‘‘1”s in each of these patterns – where h is the reporting limit. Inthe example, h is equal to two.

Using the model discovery technique described previously, themodel optimizing the cost function based on maximum Shannonentropy measurement (to minimize bias) is shown below:

Pr(R1:0 R2:0 R3:0) = 0.9997587Pr(R1:0 R2:1 R3:0) = 0.0000007Pr(R1:1 R2:0 R3:0) = 0.000043Pr(R1:1 R2:0 R3:1) = 0.0001976All other probability terms being zero

Repeating the above process for M(G2R0) through M(G6R0), re-sults in Table 4 are obtained.

Suppose during the deployment the rules R1 (Potential SSHScan) and R3 (Oversize Request – urid) are triggered, we are inter-ested in conducting a forensic inference using the information inthe models M(G1R0) through M(G6R0). In other words, we are inter-ested in identifying event patterns that satisfy the followingconditions:

when k = 1 and 3,

(1) ArgMaxRjM(GiR0),RjRk[Pr(M(GiR0)|Se:Rk = 1)] where i = 1 . . . ,6(2) Pr(arg|Se:Rk = 1)/Pr(arg) P 1

Computer A SAMBA Domain controllerVote collector + polling reporter application

Computer B Computer C Ethereal Traffic capture TCPreplay

Cwhere arg = ArgMax RjM(GiR0),RjRk[Pr(M(GiR0)|Se:Rk = 1)], i =

1 . . . ,6Condition (2) shown above basically imposes a requirement on

the admissibility of the model to participate on contributing the

UN

Table 6Voting result given R1:1.

Observation source Model source(s) Event # of votes

R1:1 R3:1 M(G1R0), M(G5R0) R2:0 2R1:1 R3:1 M(G3R0), M(G4R0), M(G6R0) R5:1 3R1:1 R3:1 M(G2R0), M(G4R0), M(G5R0), M(G6R0) R4:0 4

Table 7Result of forensic inference.

Observation source New event discovery Forensic analysis result

R1:1 R3:1 R2:0 R4:0 R5:1 R1:1 R2:0 R3:1 R4:0 R5:1

Please cite this article in press as: B.K. Sy, Integrating intrusion alert(2009), doi:10.1016/j.inffus.2009.01.001

ED

PR

OO

F

best explanation only if the model covers the observable triggeringrule(s). Event patterns that satisfy the conditions are shown in Ta-ble 5.

When R1:1 and R3:1 are observed, the best explanations fromall six models – (M(G1R0) through M(G6R0)) – satisfy all conditions;thus contributing to new information for participating in themajority voting scheme. (R2:0) from M(G1R0), (R4:0) from M(G2R0),(R5:1) from M(G3R0), (R4:0 R5:1) from M(G4R0), (R2:0 R4:0) fromM(G5R0), and (R4:0 R5:1) from M(G6R0) are six pieces of new infor-mation uncovered during the forensic inference process. Usingthese six pieces of new information to construct majority votingresult, Table 6 is obtained.

As mentioned previously, the queue size is two. Given theobservation of R1 and R3 being triggered, we will seek for the eventfrom Table 6 that has the largest number of votes with an instan-tiation value of 1. In this case, R5:1 is the winner. By observing R1and R3 in the intrusion alert, the forensic inference will furthersuggest to investigate the activities that may also have triggeredR5 (Table 7).

10. Experimental study

An experimental study was conducted to demonstrate the real-ization of AIDF in a multiple sensor environment and its ability onextracting forensic explanations based on the intrusion alerts ofmultiple sensors. The diagram in Fig. 1 depicts the set up for theexperimental simulation.

As shown in Fig. 1, the IDS sensors are placed behind a simpleNAT firewall environment. Ideally, this experiment should be con-ducted in the public network as opposed to the virtual private net-work. Due to the security policy restriction in our Universityenvironment, the experiment can only be conducted in a restrictedvirtual private network with simple NAT firewall enabled. As aconsequence, should a DoS attack occur, the NAT firewall wouldhave filtered the attack packets that target at a large number ofclosed ports. Therefore, a large number of signature rules thatshould have been triggered would not appear in an intrusion alert.

Computer 3 M1,1 … M1,4

Alert log IDS sensor 1 RL=3, QS=8 Signature Rule subset 1

Computer 9 M2,1 … M2,4

Alert log IDS sensor 2 RL=3, QS=8 Signature Rule subset 2

Computer 12 M4,1 … M4,4

Alert log IDS sensor 4 RL=3, QS=8 Signature Rule subset 4

NAT 192.168.123.254

192.168.123.0

…… …

SAMBA server

SAMBA server

SAMBA server

Signature rule set

Fig. 1. Network environment for the experimental simulation.

information to aid forensic explanation: An ..., Informat. Fusion

Page 11: Integrating intrusion alert information to aid forensic ... document… · PROOF 1 2 Integrating intrusion alert information to aid forensic explanation: An 3 analytical intrusion

T

1096

1097

1098

1099

1100

1101

1102

1103

1104

1105

1106

1107

1108

1109

1110

1111

1112

1113

1114

1115

1116

1117

1118

1119

1120

1121

1122

1123

1124

1125

1126

1127

1128

1129

1130

1131

11321133

1134

1135

1136

1137

1138

1139

1140

1141

1142

1143

1144

1145

1146

1147

1148

1149

1150

1151

1152

1153

1154

1155

1156

1157

1158

1159

1160

1161

1162

1163

1164

1165

1166

1167

1168

1169

1170

1171

1172

1173

1174

1175

1176

1177

1178

1179

1180

1181

1182

1183

1184

1185

1186

1187

1188

1189

1190

1191

1192

1193

1194

1195

1196

1197

1198

1199

1200

1201

1202

1203

1204

1205

1206

1207

1208

1209

1210

1211

1212

1213

1214

1215

1216

1217

1218

1219

1220

1221

1222

1223

1224

1225

B.K. Sy / Information Fusion xxx (2009) xxx–xxx 11

INFFUS 338 No. of Pages 17, Model 5G

16 February 2009 Disk UsedARTICLE IN PRESS

UN

CO

RR

EC

In other words, only attack packets targeting at the open serverports could pass through to trigger a limited number of signaturerules. Because of this, we refer the study as a ‘‘simulatedexperiment.”

The environment of the simulated experiment consists of a doz-en of networked computers behind a NAT router serving as a gate-way (192.168.123.254) for the network 192.168.123.0. Among thedozen of computers, four are Snort IDS sensors for the network. Atthe time of the experiment, about 3000 signature rules were avail-able. For each IDS sensor, a default setting of a queue size = 8 and areporting limit = 3 were used.

In this simulated experiment, each one of the four IDS sensors isloaded with a database consisting of 1200 unique signature rules.33% of the signature rules are overlapped between a pair of theIDS sensors. The duration of this experiment is one-month periodbetween the end of December 2005 and the end of January 2006.12 unique signatures were reported. The complete detail on thedistribution of the signature rules appeared in the alert log as re-ported by each IDS sensor can be found in the online appendix Ta-ble 1 [32]. Below is the list of the 12 unique signature rulesappeared in the alert log:

1 BLEEDING-EDGE Potential SSH Scan2 BLEEDING-EDGE VIRUS HTTP Challenge/ResponseAuthentication3 (http_inspect) OVERSIZE REQUEST-URI DIRECTORY5 BLEEDING-EDGE EXPLOIT Awstats Remote Code ExecutionAttempt6 BLEEDING-EDGE EXPLOIT XML-RPC for PHP Remote CodeInjection8 BLEEDING-EDGE Web Proxy GET Request21 (portscan) TCP Portsweep46 (http_inspect) BARE BYTE UNICODE ENCODING47 BLEEDING-EDGE Exploit Suspected PHP Injection Attack48 BLEEDING-EDGE EXPLOIT WEB PHP remote file includeexploit attempt49 BLEEDING-EDGE Proxy CONNECT Request50 BLEEDING-EDGE THCIISLame IIS SSL Exploit Attempt

As noted, only 12 unique signature rules were triggered eventhe size of the entire signature database is about 3000. Given therelatively small size of the unique signature rules being triggered,it might be reasonable to consider data mining approach such asassociation rule discovery in form of Pr(X|Y) (where X is a possibleset of ‘‘missing” rules and Y is a set of observable rules). However,while the approach based on mining the association rules is a log-ical extension of this approach, it typically also requires additional/stronger conditions. For example, association rule mining approachcould rely on the same statistical information to determine the de-gree of association as measured by not just support measure (de-fined by joint event likelihood), but also additional criteria suchas confidence measure, lift measure, leverage, or level of interest-ingness (as measured by mutual information measure)[24,35,36]. If one reduces association rule discovery to base on onlysupport measure with a priori derived from simple frequencycount, then our approach could be considered as a kind of associa-tion rule discovery. With stronger conditions as typically requiredin association rule discovery, the key to the successful applicationof association rule mining to tackle the same IDS problem will beon identifying additional/stronger conditions that would acceptassociation rules leading to improving the false positive rate with-out worsening the false negative rate. In this research, we only relyon posterior Bayesian inference.

Furthermore, during the progress of this paper an issue re-occurring a number of times is the contention on the reasonable-ness of the dozen of intrusion alerts. Specifically, is it reasonable

Please cite this article in press as: B.K. Sy, Integrating intrusion alert(2009), doi:10.1016/j.inffus.2009.01.001

ED

PR

OO

F

to expect only a dozen of alerts in the one-month period coveredby this experimental study? Unfortunately the expected numberof intrusion alerts is more complicated than just a function of time.

We assert that successful deployment of IDS requires an on-going tuning process. Initially an IDS is not uncommon to producean excessive number of false positives or false negatives dependingupon the choice of the rule set. Overtime the rule set should be ad-justed to improve the false alarms; e.g., administratively prohib-ited ICMP alert is a ‘‘famous” false alarm among certain networkinfrastructure set up involving Cisco Pix device and DNS server.Until the IDS is tuned to minimize the false alarms, the IDS alertsmay be too noisy, and may not be meaningful in regard to offeringinsights into the weakness of a current set up and how a maliciousattack could penetrate through a security zone. In other words, thereasonableness of the number of intrusion alerts in relation to theexpected number of intrusion alerts is a difficult question becauseof its time varying nature. Nonetheless, there are at least twoimportant aspects to consider for answering this difficult question:(i) the current state of security zone refinement, and (ii) the envi-ronment/source from which it is based on for deriving the ex-pected number of intrusion alerts. With respect to the firstaspect, the number of observable alerts from any IDS set up thatis still under refinement and entails an unacceptable false positiveor false negative rate would not be appropriate for establishing abenchmark for comparison purpose. With respect to the second as-pect, the number of intrusion alerts observed in one environmentis not necessarily appropriate to be used as a reference for compar-ison if the inherent environment on the kind of server services(thus the opening ports) is not at least similar. As such, we considerour finding on the size of unique signature rule set over the one-month period of the experimental study as ‘‘best effort”.

10.1. Experiment procedure

The one-month intrusion alert data set was used to create seventraining and testing data sets in the running periods. Each data setconsists of one week training and one week testing, with half of aweek overlapping between two successive data sets. In addition,the testing period is immediately right after the training period.The specific date intervals could be found in the online appendixTable 2 [32].

In the experimental study, the training data set was the signa-ture rules reported in the intrusion alerts by each sensor. These sig-nature rules were used to define the scope of the probabilisticinference models to be built for each sensor. In general, if n signa-ture rules were reported and the scope of a model is to focus onfew rules, let us say k, there are C(n,k) numbers of possible modelsfor each sensor. However, in this simulated experiment we focuson only the models that satisfy the following condition:

On any given test period, a model for this simulated experimentmust include at least one signature rule reported in some intrusionalert(s) as appeared in the training data set, and at least one signa-ture rule reported in some intrusion alert(s) as appeared in thetesting data set.

The reason for the above condition is that if a model does notinclude at least one signature rule reported in some intrusionalert(s) as appeared in the testing data set, the model is irrelevantin regard to gaining insights into answering a fundamental ques-tion related to the objective of this simulated experiment; namely,how effective is the AIDF framework and the voting mechanismdescribed in the previous sections?

It is also important to point out that the scope of the rule cov-erage in the probability model discovery process is limited by thenumber of rules triggered during the training period. Even if eachIDS sensor is loaded with thousands of rules, the size of the ruledatabase of an IDS sensor only increases the likelihood of ‘‘catch-

information to aid forensic explanation: An ..., Informat. Fusion

Page 12: Integrating intrusion alert information to aid forensic ... document… · PROOF 1 2 Integrating intrusion alert information to aid forensic explanation: An 3 analytical intrusion

T

1226

1227

1228

1229

1230

1231

1232

1233

1234

1235

1236

1237

1238

1239

1240

1241

1242

1243

1244

1245

1246

1247

1248

1249

1250

1251

1252

1253

1254

1255

1256

1257

1258

1259

1260

1261

1262

1263

1264

1265

1266

1267

1268

1269

1270

1271

1272

1273

1274

1275

1276

1277

1278

1279

1280

1281

1282

1283

1284

1285

1286

1287

1288

1289

1290

1291

1292

1293

1294

1295

1296

1297

1298

1299

1300

1301

1302

1303

1304

1305

1306

1307

1308

1309

1310

1311

1312

1313

1314

1315

1316

1317

1318

1319

1320

1321

1322

1323

1324

1325

12 B.K. Sy / Information Fusion xxx (2009) xxx–xxx

INFFUS 338 No. of Pages 17, Model 5G

16 February 2009 Disk UsedARTICLE IN PRESS

CO

RR

EC

ing” an attack. If during the test period only a dozen of rules trig-gered, the scope coverage of the probability model could only beat most a dozen of rules as opposed to the number of the rulesin the database. It is because probability model discovery is onlyapplied to cover the scope of the rules triggered. In this regard,one must be mindful that extracting forensic explanation fromintrusion alerts is opportunistic.

For the purpose of this simulated experiment, four models werebuilt for each IDS sensor using only the training data in a given testperiod. Each model (satisfying the conditions above) consists ofthree signature rules that have been reported. The distribution ofthe rule occurrence patterns that have at least one triggered ruleis shown in the online appendix Table 3 [32]. This distributioninformation is then used for formulating constraints for probabilitymodel discovery.

Probability constraints based on the frequency count informa-tion were formulated for the discovery of inference models of eachIDS sensor of each test. An example of the formulation of the prob-ability constraints is shown in Appendix 1. This example uses thedata from the test set 3, IDS sensor 2, and covers the scope involv-ing R1, R47, and R48 (shown in the online appendix Table 4 [32]).As shown in this example, the probability constraints result in anunder-determined algebraic system. The illustration shown inAppendix 1 presents three of the multiple solutions of the under-determined algebraic system. Whenever multiple solutions maybe available as a result of an under-determined algebraic system,the optimal solution is selected as the probability model for infer-ring the most probable forensic explanation, where the optimiza-tion criterion is the minimal bias as defined mathematicallyusing Shannon entropy.

For comparison purpose, the same constraint set is also used toderive a naïve Bayes model under the assumption of independence.In other words, the model discovery process is applied to obtainthe marginal probability information. The joint probability infor-mation is then derived from the product of the marginal probabil-ities. Naïve Bayes is supposed to offer computational advantage.Therefore, the basic idea is to approximate the joint probabilitymodel by a naïve Bayes model. This is to establish a benchmark,and to speculate whether the probabilistic inference using the naïveBayes model could achieve a result similar to the joint probabilitymodel discovered in the AIDF.

During the test period, intrusion alerts being reported wereused as the testing data for the evaluation purpose. Recall thatthe default reporting/log limit is three at the data collection phrase.Based on the default setting, there are two possibilities of the realworld situation. First, three rules are reported in an alert. Second,less than three rules are reported in an alert. An alert will never re-port more than three rules because reporting/log limit is three.

When three rules are reported in an alert, there could be morethan three rules being triggered but not reported. Therefore, whenan alert reports three rules, one will not be able to tell in reality

UN

Table 8Snapshot of experimental result.

January 7–13 (Train) Jan 14-20 (Test) T2

IDS1 Modelbelief

IDS2

8,6,5, (2,3),1, (47,48),49 (mask 3),2 M11(2,3,5) 111 M21(2,M12(2,3,6) 111 M22(1,M13(2,5,6) 100 M23(1,M14(3,5,6) N/A M24(1,

Forensichypothesis

111/ 100 Forenshypoth

Forensic explanation R1:0 R2:1 R3:1 R47:0R48:0

Please cite this article in press as: B.K. Sy, Integrating intrusion alert(2009), doi:10.1016/j.inffus.2009.01.001

ED

PR

OO

F

how many rules are triggered. If the exact number of triggeredrules is not known in an alert, one could not use it to establish a‘‘ground truth” for an experimental evaluation. As such, only a sub-set of the real world alerts qualifies as the ‘‘ground truth” for anexperimental evaluation; i.e., alerts that report less than threerules.

The subset constituting the ground truth guarantees the ob-served real world alerts to have no hidden unreported rules. Thissubset of alerts is used to derive test cases by masking triggeredrules away. Since an alert in the ground truth has less than threerules, it could have at most two triggered rules. In the alerts wheretwo rules were reported, each rule was masked one-at-a-time togenerate a test case. We do not mask both rules away at the sametime because this is equivalent to have no alert (no rule gets trig-gered). To be precise, even though the real world alerts are cap-tured using the default setting with a log limit = 3, only thesubset constituting the ground truth is useful for the experimentalevaluation, and the evaluation is (and can be) based only on a sce-nario where at most two rules are triggered, or an equivalence of a‘‘simulated” log limit = 2.

During the test phase, the unmasked rule in the alert of a testcase is used as an evidence/observation in the probabilistic infer-ence. The outcome of the probabilistic inference is a ‘‘belief state-ment.” This belief statement produced by each of the four modelsof each sensor was then used to construct the ‘‘votes” for generat-ing a ‘‘forensic hypothesis.” The forensic hypothesis of each sensoris then used to construct the ‘‘votes” for generating a forensicexplanation. Ideally, the forensic explanation should contain nofurther claim other than the signature rule that has already beenreported and that was masked. Otherwise, it results in a false posi-tive in the forensic explanation. Similarly, if the forensic explana-tion fails to discover the masked rules, it results in a false negative.

10.2. Preliminary experimental result and evaluation metrics

In the one-month period, a total of about 1.3 million packetstraversed through, and inspected by, Snort. Spanning over the test-ing periods of the experiment as shown in the online appendix Ta-ble 4 [32], there were 47 test cases originated from the 12 uniquesignatures. The complete experimental results are shown in theonline appendix Tables 4 and 5 (for AIDF) (for naïve Bayes) [32].In reference to the result of all the test cases, the information aboutthe log report limit is not provided to probabilistic inference duringthe course of generating a forensic explanation. Table 8 shows asnapshot of the experimental result extracted from the onlineappendix Table 4 [32] to facilitate a further discussion.

The snapshot shown in Table 8 is for a test case in T2 with thedata in January 7 through 13 used for training purpose. During thetraining period, five unique rules and two pairs of rules (triggeredsimultaneously) were reported in the intrusion alerts. This testcase occurred between January 14 and 20. In this test case R2 is

Modelbelief

IDS3 Modelbelief

IDS4 Modelbelief

47,48) P(se)=0 M31(3,5,49) N/A M41(8,47,48) N/A2,48) 010 M32(3,21,49) N/A M42(6,47,48) N/A2,47) 010 M33(5,49,50) N/A M43(47,48,49) N/A47,48) N/A M34(3,5,50) N/A M44(6,8,47) N/A

icesis

010 Forensichypothesis

N/A Forensichypothesis

N/A

information to aid forensic explanation: An ..., Informat. Fusion

Page 13: Integrating intrusion alert information to aid forensic ... document… · PROOF 1 2 Integrating intrusion alert information to aid forensic explanation: An 3 analytical intrusion

T

1326

1327

1328

1329

1330

1331

1332

1333

1334

1335

1336

1337

1338

1339

1340

1341

1342

1343

1344

1345

1346

1347

1348

1349

1350

1351

1352

1353

1354

1355

1356

1357

1358

1359

1360

1361

1362

1363

1364

1365

1366

1367

1368

1369

1370

1371

1372

1373

1374

1375

1376

1377

1378

1379

1380

1381

1382

1383

1384

1385

1386

1387

1388

1389

1390

1391

1392

1393

1394

1395

1396

1397

1398

1399

1400

1401

1402

1403

1404

1405

1406

1407

1408

1409

1410

1411

1412

1413

1414

1415

1416

1417

1418

1419

1420

1421

1422

1423

1424

1425

1426

1427

1428

1429

1430

1431

1432

1433

14341435

1436

1437

1438

1439

144014411442144314441445

14461447

14481449

14501451

1452

1453

14541455

1456

1457

1458

B.K. Sy / Information Fusion xxx (2009) xxx–xxx 13

INFFUS 338 No. of Pages 17, Model 5G

16 February 2009 Disk UsedARTICLE IN PRESS

UN

CO

RR

EC

provided as an evidence, whereas R3 is masked away to emulatethe side effect of reporting/log limit. As depicted above, there arefour IDS sensors (IDS1–IDS4). Four models are derived for each ofthe sensors. For IDS sensor 1, the first model M11 covers the scopecharacterized by (R2, R3, R5). Likewise, the first model of the sec-ond and third IDS sensor – M21 and M31, covers the scope charac-terized by (R2, R47, R48) and (R3, R5, R49) respectively.

One important point to note is that the scalability of our pro-posed approach does not depend on the number of rules or thecombination of (triggered) rules. Rather, the scalability of our pro-posed approach can always be made manageable through the ruleset partitions of the AIDF discussed in Section 9. Referring to Table8, there are 16 models. These 16 models collectively cover 11 rules(a combination of 211) while each model has only a representa-tional complexity in the order of 23. This complexity can remainthe same if we just increase the number of models to cover an ex-panded scope of increased number of rules.

When R2 is reported as a simulated alert, it is used by eachmodel as an evidence for the probabilistic inference whenever pos-sible. Note that the most probable explanation generated by thesecond model of the first IDS sensor M12 is (R2:1 R3:1 R6:1) – rep-resented by the string ‘‘111” in Table 8. Furthermore, the first mod-el of second IDS sensor M21 and the third IDS sensor M31 will not beutilized for the probabilistic inference. The first model of the thirdIDS sensor M31 will not be utilized for the probabilistic inferencebecause the evidence R2:1 is out of scope with respect to M31 –represented by ‘‘N/A.” Note that a model could be out of scope inone test case but will be ‘‘in scope” for at least one test case in thissimulated experiment because of the condition stated previously(under the description of experimental procedure). Also, despiteM21 covers the scope consisting of R2, Pr(R2:1) is zero. Therefore,it is not possible to conduct probabilistic inference for derivingArgMax[Pr(R47,R48|R2:1)]. Thus M21 and M31 are not utilized inconducting probabilistic inference in this case.

In the process of producing a forensic hypothesis produced bythe IDS sensor 1, there are three counts of R2:1 and two counts ofR3:1 while there is zero count for R2:0 and R3:0. In addition, thereis equal number of counts for R5:0 and R5:1, as well as R6:0 andR6:1; thus neutralizing each other. As a result, the forensic hypoth-esis resulted from IDS sensor 1 will be (R2:1, R3:1). Likewise, theforensic hypothesis resulted from IDS sensor 2 will be (R1:0, R2:1,R47:0, R48:0). The forensic explanation derived from these twoforensic hypotheses is then (R1:0, R2:1, R3:1, R47:0, R48:0).

In the snapshot shown in Table 8 one model of IDS sensor 1, twomodels of IDS sensor 2, and all models of IDS sensors 3 and 4 arenot utilized in probability inference given the case (R2:1). Ideally,it is desirable to utilize a model as often as possible. Therefore,one aspect of the evaluation would be on model utility as mea-sured by the percentage of a model being utilized for probabilisticinference with respect to the number of test cases. In other words,it measures how often a model is relevant and used in a probabilis-tic inference, as well as participating in voting.

Furthermore, it is also desirable for a model to produce an‘‘accurate” probabilistic inference result. The accuracy could bemeasured by conventional metrics using false positives and falsenegatives. However, since a model may not participate in probabi-listic inference due to either out of scope or Pr(Se) = 0, two addi-tional metrics are introduced when evaluating the probabilisticinference accuracy of a model; namely, neglected negative andfalse silent. Neglected negative refers to the scenario where a ruleis not triggered but is not considered due to out of scope. An exam-ple is R49 that appears in the training data, but not triggered in theparticular case shown in Table 8, and is not considered by M11 dueto out of scope. False silent refers to the scenario where a rule istriggered but is not considered due to out of scope. Generallyspeaking, we could consider neglected negative and false silent

Please cite this article in press as: B.K. Sy, Integrating intrusion alert(2009), doi:10.1016/j.inffus.2009.01.001

ED

PR

OO

F

as the scenario where a model remains neutral by not committingto a statement about whether a rule is triggered or not. Note thatwhen model utility is low, we could expect a relatively high per-centage of neglected negative and/or false silent.

In addition, we should be aware that the probabilistic inferencedoes not have to be accurate in order to generate an accurateforensic explanation. The snapshot shown above illustrates onesuch case. Although models M11 and M12 did not generate the cor-rect inference statement about R5 and R6, it was compensated bythe inference statement of M13 (that allows us to cancel R5:0 withR5:1, and R6:0 with R6:1). In other words, the evaluation of accu-racy should be conducted in both the model level as well as in thecase level (note that there are 47 cases in this experiment).

In summary, the evaluation will be conducted in both model le-vel and case level. In the model level, each model will be evaluatedfor its utility, and its accuracy as defined by false positive, falsenegative, false silent, and neglected negatives. In the case level,we will evaluate whether the forensic explanation for a test caseis a perfect discovery, imperfect discovery, or an unsuccessful dis-covery. A perfect discovery is the case where the masked element(rule) is successfully uncovered with no incorrect conclusion onthe other rules. Imperfect discovery is the case where the maskedelement (rule) is successfully uncovered but it also includes incor-rect conclusion about other rules. Unsuccessful discovery is thecase where it fails to uncover the masked element.

For comparison purpose, naïve Bayes reasoning was also con-ducted so that it could serve as a benchmark reference. Since thescope coverage of the naïve Bayes model is identical to that ofthe model discovery, the model utility percentage remains thesame in both cases. (It is further noted that out of scope casesdue to Pr(Se) = 0 happen to be the same.) In regard to the inferenceaccuracy and case level evaluation, the results of naïve Bayes rea-soning were tabulated side-by-side with that using the joint prob-ability model discovery method described in Section 6. This allowsus to gain insights into the validity and the effectiveness of theprobability model discovery process for AIDF.

In additional to the evaluation criteria just described, the exper-imental results are also used to derive two pieces of summaryinformation for better understanding the potential benefits of theAIDF as related to the integration of the intrusion alerts from mul-tiple IDS sensors:

1. What is the percentage of perfect discovery that is attributed tocombining alert information from multiple IDS sensors?

2. What is the percentage of cases where the scope coverage of theforensic explanations is improved upon by combining informa-tion from multiple IDS sensors rather than relying on the infor-mation from a single IDS sensor?

10.3. Summary evaluation result

1. Model level evaluation (see Tables 9–11):Criteria – model utility, inference accuracy

1a. Model utility percentage1b. Inference accuracy

2. Case level evaluation:Criterion: Discovery success rate

3. Significance of AIDF as related to information integration for aforensic inference:a. Percentage of perfect discovery that is attributed to combin-

ing alert information from multiple IDS sensors: 16.67%b. Percentage of cases where the scope coverage of the forensic

explanations is improved upon by combining informationfrom multiple IDS sensors rather than relying on the infor-mation from a single IDS sensor: 87%

Note that the results just shown are based on masking off one oranother alert for instances of multiple alert triggers because of the

information to aid forensic explanation: An ..., Informat. Fusion

Page 14: Integrating intrusion alert information to aid forensic ... document… · PROOF 1 2 Integrating intrusion alert information to aid forensic explanation: An 3 analytical intrusion

T

OF

1459

1460

1461

1462

1463

1464

1465

1466

1467

1468

1469

1470

1471

1472

1473

1474

1475

14771477

1478

1479

1480

1481

1482

1483

1484

1485

1486

1487

1488

1489

1490

1491

1492

1493

1494

1495

1496

1497

1498

1499

1500

1501

15031503

1504

1505

1506

1507

1508

1509

1510

1511

1512

1513

1514

1515

1516

1517

1518

1519

1520

1521

1522

1523

1524

1525

1526

1527

1528

1529

1530

1531

1532

1533

1534

Table 9Model utility percentage.

IDS1 IDS2 IDS3 IDS4

M1 (%) 49 38 40 19M2 (%) 49 36 32 23M3 (%) 45 30 19 19M4 (%) 45 28 30 21

Table 10Inference accuracy.

Inferred: 0 Inferred: 1 Inferred: Neutral

Actual: 0 True negative: False positive: Neglected negative:42.6% (NB 47.46%) 1.8% (NB: 0.2%) 55.6% (NB: 52.34%)

Actual: 1 False negative: True positive: False silence:7% (NB: 30.14%) 87.4% (NB: 65.75%) 5.6% (NB: 4.11%)

Table 11Discovery success rate.

Total # of cases: 47 Prob. model discovery (%) Na Bayes (%)

Perfect discovery 40 4Imperfect discovery 28 12Unsuccessful discovery 32 84

14 B.K. Sy / Information Fusion xxx (2009) xxx–xxx

INFFUS 338 No. of Pages 17, Model 5G

16 February 2009 Disk UsedARTICLE IN PRESS

UN

CO

RR

EC

need to establish a ground truth for the experimental validationwith report limit = 2. In retrospect, the experimental validationcould be improved upon to better reflect the ground truth of theinhibiting negative caused by the real traffic data if we could affordto perform multiple replays on the traffic data. Specifically, thetraffic data could be replayed multiple times by incrementallyincreasing the report limit up to the queue size, and then byrepeating again for increased queue size. In doing so, the amountof effort required for the experimental study would increase in apolynomial order with respect to the level of improvement onthe ground truth. Despite the additional burden, further experi-mental study was conducted to simulate the replay for a betterunderstanding on the possible range of the number of ‘‘true inhib-iting negative.” This is achieved by setting the reporting limit iden-tical to the size of the queue and we extrapolate the number of trueinhibiting negative based on the time that the rules are triggered.The results are shown in the graph below:

-2.5

-2

-1.5

-1

-0.5

0

1 2 3 4

# of simultaneous triggered rules

Lo

g s

cale

of

the p

erc

en

tag

e o

f

sim

ult

an

eo

us t

rig

gere

d r

ule

s

Series1

The graph above shows a log-linear relationship between thenumber of rules triggered simultaneously as related to causinginhibiting negative and their relative percentage. While thegraph exhibits a log-linear relationship, readers are cautionedfor not rushing into a conclusion on power law distribution.In this study we do not have sufficient data to conduct a statis-tical hypothesis test to validate/invalidate whether the distribu-tion of the percentage of simultaneous triggered rules followspower law property with an underlying binomial or Poissondistribution.

Yet another aspect that we consider is the pre-condition for theoccurrence of inhibiting negative; i.e., inhibiting negative could onlyoccur if packets or packet streams could trigger multiple rulessimultaneously. In other words, packet streams that will not triggermultiple rules cannot cause inhibitive negative. Consequently theexistence of packet streams that could trigger multiple rules oversome pre-defined unit time may be an indication of increased like-lihood of the occurrence of inhibiting negative. To gain further in-sight into the likelihood of the occurrence of inhibiting negative,we also analyzed the number of rules triggered by the (attack) trafficover short term windows (i.e., time intervals in seconds). An analy-sis of the traffic data accumulated from our network shows that the

Please cite this article in press as: B.K. Sy, Integrating intrusion alert(2009), doi:10.1016/j.inffus.2009.01.001

ED

PR

Onumber of unique rules triggered over different sizes of short termwindow indicates some piece-wise log-linear relationship:

Rule triggering interval percentage

-5

-4

-3

-2

-1

01 2 3 4 5 6 7 8 9 10 11 12 13 14 15

Time interval by sec

Lo

g s

cale

of

trig

ger

ed r

ule

per

cen

tag

e

Series1

10.4. Discussion

The model utility evaluation result shows that the utility of themodels constructed for this experiment ranges from 19% to 49%.Recall that a model covers only a very small subset of all the avail-able signature rules of an IDS sensor. The choice of signature ruleset partition for an IDS sensor, and the selection of small subsetof rules for modeling, have a direct impact on the model utility per-centage. Unless historical network traffic bears a predictable rela-tionship with the future network traffic for estimating thedistribution of the triggered rule set, it remains as an open ques-tion as to the choice of signature rule set partition for an IDS sensorand the selection of small subset of rules for optimizing the modelutility percentage.

In reference to the inference accuracy, the results are shown inTable 10. The performance of the naïve Bayes is shown in thebracket labeled as ‘‘NB.” It is noted that the proposed method sig-nificantly outperforms naïve Bayes in terms of true positive andfalse negative categories. The two methods perform relatively sim-ilar to each other in the other categories. While the proposed ap-proach achieves an encouraging high accuracy on the truepositive, it is also noted that the choice of the rule coverage by amodel and the number of models chosen for each IDS sensor tendsto be too conservative in lieu of a relatively high percentage of ne-glected negative – suggesting that the proposed approach tends toremain neutral and refrain from contributing towards the genera-tion of a forensic explanation on rules that are not actually trig-gered. Although it is conceptually conceivable to improve theneglected negative and false silence rate by twinkling the modelutility factor via different choices of scope coverage, the redistribu-tion of the rules for the models, however, does not guarantee theoutcomes to fall into the categories of true positive or true nega-tive. Furthermore, it remains as a challenge in a real world deploy-

information to aid forensic explanation: An ..., Informat. Fusion

Page 15: Integrating intrusion alert information to aid forensic ... document… · PROOF 1 2 Integrating intrusion alert information to aid forensic explanation: An 3 analytical intrusion

T

1535

1536

1537

1538

1539

1540

1541

1542

1543

1544

1545

1546

1547

1548

1549

1550

1551

1552

1553

1554

1555

1556

1557

1558

1559

1560

1561

1562

1563

1564

1565

1566

1567

1568

1569

1570

1571

1572

1573

1574

1575

1576

1577

1578

1579

1580

1581

1582

1583

1584

1585

1586

1587

1588

1589

1590

1591

1592

1593

1594

1595

15961597

1598

1599

1600

1601

1602

1603

1604

1605

1606

1607

1608

1609

1610

1611

1612

1613

1614

1615

1616

1617

1618

1619

1620

1621

1622

1623

1624

1625

1626

1627

1628

1629

1630

1631

1632

1633

1634

1635

1636

1637

1638

1639

1640

1641

1642

1643

1644

1645

1646

1647

1648

1649

1650

1651

1652

1653

1654

1655

1656

1657

1658

1659

1660

1661

B.K. Sy / Information Fusion xxx (2009) xxx–xxx 15

INFFUS 338 No. of Pages 17, Model 5G

16 February 2009 Disk UsedARTICLE IN PRESS

UN

CO

RR

EC

ment environment to know whether it is a true negative when arule is not triggered, except the potential inhibiting negative in cer-tain cases.

In reference to the case level evaluation, the discovery rate isvery encouraging when the results of AIDF are compared againstthat of naïve Bayes. As discussed previously, one should expectthat naïve Bayes would not perform as well because of the natureof the problem. Note that extracting forensic explanation fromintrusion alerts takes advantage on the possible association amongthe rules that do not just co-occur independently. Yet the underly-ing independence assumption of the naïve Bayes will naturallyignore the useful co-occurrence information. This is confirmed bythe experimental result that shows a significantly lower true posi-tive rate. Although theoretically naïve Bayes has an advantage interms of computational simplicity and efficiency, such an advan-tage is marginal given the size of the problem and the consider-ation of the model discovery complexity that involves inequalityconstraint.

Imperfect discovery rate, however, seems high even though it isstill better than that of the naïve Bayes. It is because the imperfectdiscovery rate accounts for about 40% of the successful discovery ofthe hidden rule. There are two aspects that could be further inves-tigated for improving the imperfect discovery rate. Note thatimperfect discovery originates from the cases that include partiallyincorrect statement on the occurrence of the irrelevant rules inaddition to the correct claim on the masked element (rule). Thereare three open questions that could be further investigated forimproving the imperfect discovery rate. First, the success on gener-ating forensic explanation depends on the ability of reasoning theaccurate mathematical constraints for model discovery. How dowe automate the process for generating the optimal constraintset? Second, alternative voting strategy other than the ‘‘equal-weight” strategy may help to cancel the effect of incorrect votingbias. What domain knowledge could be used to devise voting strat-egy that improves the imperfect and unsuccessful discovery in theprocess of generating a forensic explanation? Third, if operationalknowledge (based on, for example, security policy and firewall ruleimplementation) on the maximum number of triggering conditionsis known a priori, one could carefully select a queue size matchingthe maximum number of rules that could be triggered simulta-neously, thus offering additional insights into improving inferenceaccuracy and imperfect discovery rate. In other words, could onederive the relationship between the choice of queue size and a gi-ven rule set when operational knowledge is available?

In regard to the potential contribution of AIDF to integratinginformation from multiple IDS, and particularly in the area offorensic inference based on intrusion alerts, the results are encour-aging. 16.67% of successful (perfect) discovery of hidden signaturerules would not be possible without integrating intrusion alertsfrom multiple IDS sensors. Furthermore, in 87% of the test casesthe scope coverage of the forensic explanations is improved uponby combining the information about the intrusion alerts producedby the multiple sensors.

In summary, although the experimental results demonstratethe success on the analytical intrusion detection framework anddistributive intrusion detection, the following five factors bearinteresting research open questions for further investigation:

Choice of signature rule set partitionNumber of models per IDSScope of rule coverage for each modelVoting strategyQueue size with respect to operational knowledge

Despite the open research questions just raised, perhaps themost valuable real world forensic inference that we have learned

Please cite this article in press as: B.K. Sy, Integrating intrusion alert(2009), doi:10.1016/j.inffus.2009.01.001

ED

PR

OO

F

from the one-month experimental study is a common attack onour site targeting at the exploit of the weakness of PHP and Perlapplications that provide web interaction and database accesswithout performing input validation – a vulnerability that couldbe exploited for remote arbitrary code execution. This explanationhas an equivalent semantic interpretation of the conjunctive com-bination of (R2:1 R3:1), (R5:1), (R6:1), (R5:1 R6:1) and (R47:1R48:1). In particular, this explanation is derived based on the dis-covery of three consistent events shown in Table 4 of the onlineappendix [28]: (a) signature rules 2 and 3 co-occurred throughoutthe one-month period, (b) rules 5 and 6 occurred separately andjointly in half of the one-month period, as well as co-occurred to-gether with rules 2 and 3, and (c) rules 47 and 48 co-occurred afterthe co-occurrence of rules 2 and 3.

11. Conclusion

It is rare for an intrusion detection system to be operated in fulllogging mode for the purpose of forensic data gathering because ofthe practical consideration and implementation of anti-DoS strat-egy in a real world deployment. Consequently, there is a gap be-tween the information useful for forensic and that revealed byintrusion detection. This paper presents a novel approach for gen-erating the most probable forensic explanation from intrusionalerts and brings forth two specific contributions. First, a probabi-listic inference approach is proposed for uncovering hidden infor-mation due to, for example, inhibiting negative – thus bridgingintrusion detection with forensic analysis. Second, the proposedapproach is generalized for an analytical intrusion detectionframework for realizing distributive intrusion detection with agranularity at the level of rule partition, and a voting mechanismfor integrating intrusion alerts from different IDS sensors distrib-uted cross the same subnet or different subnets.

To demonstrate the relevancy of this research to the state-of-the-art intrusion detection technology, we investigated the under-lying implementation of rule inspection engine in Snort. Currentimplementation of the rule inspection engine for intrusion detec-tion in Snort reports only a partial list of all matching rules due tothe consideration of the anti-DoS strategy and the need of real timeperformance. We show how our proposed model discovery andprobabilistic inference mechanism could be applied to uncoveringthe hidden matching rules for generating forensic explanationsfrom intrusion alerts. A preliminary experimental study is con-ducted to better understand its practicability and scalability. Theresults of this proposed approach is compared favorably againstnaïve Bayes approach. In our future research, we plan to furtherinvestigate the open questions raised from the five factors discov-ered in our experimental study. These open questions bear an im-pact on further improving the effectiveness and usefulness of theproposed framework for (i) realizing distributive IDS, and (ii) inte-grating intrusion alerts from different IDS sensor sources for thepurpose of forensic inference and generating forensic explanation.

Acknowledgements

The author is indebted to the reviewers for many useful sugges-tions. This research is supported in part under the grants from PSC-CUNY Research and Queens College/CUNY Graduate InitiativeInvestment. Some initial findings of this paper appeared in the2006 International Workshop on the Application of Data Miningto Security Control [33]. The author is grateful to Professor GiorgioGiacinto for his guidance on the emphasis of the paper suitable forthis special issue, and the discussion and comments offered by Pro-fessors Soon Chun of the College of Staten Island, Douglas Salane ofJohn Jay College, and Jun Zheng of Queens College.

information to aid forensic explanation: An ..., Informat. Fusion

Page 16: Integrating intrusion alert information to aid forensic ... document… · PROOF 1 2 Integrating intrusion alert information to aid forensic explanation: An 3 analytical intrusion

1662

1663

1664

1665

1666

1667

1668

1669

1670

1671

1672

1673

1674

1675

1676

1677

1678

1679

168116821683

168416851686

168716881689

169016911692

169316941695

169616971698

1699

1700

1701

170317041705170617071708170917101711

171217131714171517161717

171817191720172117221723

172417251726172717281729

173017311732173317341735

173617371738173917401741

174217431744174517461747

174817491750175117521753

175417551756175717581759

1760

1761

1762

1763

1764

1765

1766

1767

1768

1769

1770

1771

1772

1773

1774

1775

1776

1777

1778

1779

1780

1781

1782

1783

1784

16 B.K. Sy / Information Fusion xxx (2009) xxx–xxx

INFFUS 338 No. of Pages 17, Model 5G

16 February 2009 Disk UsedARTICLE IN PRESS

Appendix 1

Data source: online appendix Table 4 – Period 3, IDS2.Model scope coverage: (14748), log/report limit = 2, data size

for period 3 = 152,399.Data size of rule occurrence pattern (R1:0 R2:0 R21:0 R46:0

R47:0 R48:0) = 152,234.Probability constraint formulation:

C1: (143 + 15)/152399 = 0.00103675 P Pr(R1:1).C2: Pr(R1:1) P 143/152399 = 0.000938.C3: Pr(R47:1, R48:1) = 15/152399 = 0.000098.C4: Pr(R1:0 R47:1 R48:0) = 6/152399 = 0.000039.C5: Pr(R1:0 R47:0 R48:0) = (1 + 152234)/152399 = 0.99892.

C6:P

R1,R47,R48 Pr(R1 R47 R48) = 1.

1785

1786

1787

Denote Pr(R1:0 R47:0 R48:0) by P0, Pr(R1:0 R47:0 R48:1) byP1, . . . , Pr(R1:1 R47:1 R48:1) by P7, the constraints C1–C5 couldbe reformulated as below:

1788

C1:

(Row 1) M(Row 2) Z(Row 3) Z(Row 4) Z(Row 5) Z(Row 6) Z(Row 7) Z

Please ci(2009), d

P4 + P5 + P6 + P7 + S1

NO

aximize Z0 = �2 + 2P0 + 2P1 + 2P2 + 3P3 + 20 = 0.0051 = 0.002 �P2 �P3 �P6

2 = 0.0033 = 04 = 0.99 �P0

5 = 1 �P0 �P1 �P2

te this article in press as: B.K. Sy, Integratoi:10.1016/j.inffus.2009.01.001

= 0.00103675

C2: P4 + P5 + P6 + P7 � S2 = 0.000938 C3: P3 + P7 = 0.000098 C4: P2 = 0.000039 C5: P0 = 0.99892 C6: P0 + P1 + P2 + P3 + P4 + P5 + P6 + P7 = 1

17901790

CTED

17911792

P4 + 3P5 + 4P6 + 5P7 � S0 � S1 � S2

�P4 �P5 �P6 �P7 +S0

�P7 +S1

�P1 �P3 �P5 �P7 +S2

�P6 �P7

�P3 �P4 �P5 �P6 �P7

EBelow shows three model solutions by solving the above under-determined algebraic system using SVD decomposition:

17941795179617971798179918001801180218031804180518061807

R1 R47 R48

18081809181018111812181318141815181618171818181918201821

M1: Pr(R1 R47R48)

M2: Pr(R1 R47R48)

R(Optimal) M3: Pr(R1R47 R48)

18221823182418251826182718281829183018311832183318341835

0 0 0 0.99892 0.99892 0.99892 18361837183818391840184118421843184418451846184718481849 0 0 1 1.03E�04 1.3E�05 R1.3E�05 18501851185218531854185518561857185818591860186118621863 0 1 0 3.9E�05 3.9E�05 3.9E�05 18641865186618671868186918701871187218731874187518761877 0 1 1 0 0 2.00E�05 18781879188018811882188318841885188618871888188918901891 1 0 0 8.40E�04 8.40E�04 8.40E�04 18921893189418951896189718981899190019011902190319041905 1 0 1 0 0.00002 0.00007 19061907190819091910191119121913191419151916191719181919 1 1 0 0 C0.00007 0.00002 1920 1 1 1 9.80E�05 9.80E�05 7.80E�05

1921

1922

1923

1924

1925

1926

1927

1928

1929

1930

1931

1932

1933

1934

UAppendix 2

In this appendix the example shown in Section 6 will be used toillustrate the application of the Kuenzi, Tzschach, and Zehnder(KTZ) approach to derive an initial feasible solution. The basic ideabehind the KTZ approach for solving an algebraic system of linearconstraints is to reformulate the constraint set by introducingadditional variables – one for each constraint. Using the examplein Section 5 (and the algebraic formulation shown in Section 6),

Z0 ¼ 0:005� P4 � P5 � P6 � P7 þ S0.

ing intrusion alert

PR

OO

F

Z1 ¼ 0:002� P2 � P3 � P6 � P7 þ S1.Z2 ¼ 0:003� P1 � P3 � P5 � P7 þ S2.Z3 ¼ �P6 � P7.Z4 ¼ 0:99� P0.Z5 ¼ 1� P0 � P1 � P2 � P3 � P4 � P5 � P6 � P7.Zi 6 0; i ¼ 0; . . . ;5.

The above (in)equalities can be thought of as a constraint set ofan optimization problem with a cost function Min[Z0 + Z1 +Z2 + Z3 + Z4 + Z5]. Note that a feasible solution of this new optimiza-tion problem is a vector of 17 parameters [Z0 Z1 Z2 Z3 Z4 Z5 P0 P1 P2

P3 P4 P5 P6 P7 S0 S1 S2]. If the global minimum can be achieved inthis new optimization problem, this is equivalent toZ0 = Z1 = � � � = Z5 = 0, which in turn gives a feasible solution satisfy-ing the non-negativity requirement for the original problem. Thatis, Pis in the global optimal solution [0 0 0 0 0 0 P0 P1 P2 P3 P4 P5

P6 P7 S0 S1 S2] is a non-negative feasible solution to the originalmodel discovery problem.

To illustrate the mechanism behind the method by Kuenzi et al.,the example shown above is reformulated in tableau form below:

Minimize Z ¼ Z0 þ Z1 þ Z2 þ Z3 þ Z4 þ Z5 ¼X

i

Zi

$Minimize Z ¼ 2� 2P0 � 2P1 � 2P2 � 3P3 � 2P4 � 3P5

� 4P6 � 5P7 þ S0 þ S1 þ S2

$Maximize� Z ¼ Z0 ¼ �2þ 2P0 þ 2P1 þ 2P2 þ 3P3 þ 2P4

þ 3P5 þ 4P6 þ 5P7 � S0 � S1 � S2

inf

C1

orm

C2

ation to a

C3

id fo

C4

ren

C5

sic e

C6

xpla

C7

natio

C8

n: A

C9

n ...

C10

, Info

C11

rmat

C12

. Fusi

C13

Constant P0 P1 P2 P3 P4 P5 P6 P7 S0 S1 S2

R1

Z0 �2 2 2 2 3 2 3 4 5 �1 �1 �1 R2 Z0 0.005 �1 �1 �1 �1 1 R3 Z1 0.002 �1 �1 �1 �1 1 R4 Z2 0.003 �1 �1 �1 �1 1 R5 Z3 0 �1 �1 R6 Z4 0.99 �1 R7 Z5 1 �1 �1 �1 �1 �1 �1 �1 �1

Examine Z0 or row R1 to determine the entry/entries below a posi-tive Z0-row that is/are negative. In this case, we have (R2 C7), (R2C8), (R2 C9), (R2 C10), (R3 C5), (R3 C6), (R3 C9), (R3 R10), (R4 C4),(R4 C6), (R4 C8), (R4 C10), (R5 C9), (R5 C10), (R6 C3), (R7 C3) through(R7 C10).

Note that altogether there are 14 variables plus three slack var-iable (P0, . . . , P7, Z0, . . . , Z5,S0,S1,S2) and six equations (rows 2–7) notincluding the cost function (row 1). It is always possible to set thevariables appearing at the Z0-row (row 1) to be zero while solvingthe equations in rows 2–7 for the remaining variables. For exam-ple, in the tableau shown above, P0, . . . , P7, S0, S1, and S2 appear atZ0-row (row 1). If P0 = � � � = P7 = S0 = � � � = S2 = 0, equations on rows2–7 can be used to solve for Z0, through, Z5. In such a case, the valuefor Z0, through Z5 can readily be read off from the ‘‘constant” col-

on

Page 17: Integrating intrusion alert information to aid forensic ... document… · PROOF 1 2 Integrating intrusion alert information to aid forensic explanation: An 3 analytical intrusion

1935

1936

1937

1938

1939

1940

1941

1942

1943

1944

1945

1946

1947

1948

1949

1950

1951

1952

1953

1954

1955

1956

19581959196019611962196319641965196619671968196919701971

19721973197419751976197719781979198019811982198319841985

19861987198819891990199119921993199419951996199719981999

20002001200220032004200520062007200820092010201120122013

20142015201620172018201920202021202220232024202520262027

20282029203020312032203320342035203620372038203920402041

20422043204420452046204720482049205020512052205320542055

20562057205820592060206120622063206420652066206720682069

20702071207220732074207520762077207820792080208120822083

2084

2085

2086

2087

2088

2089

2090

209120922093209420952096209720982099210021012102

21032104210521062107210821092110211121122113211421152116211721182119212021212122212321242125212621272128212921302131

B.K. Sy / Information Fusion xxx (2009) xxx–xxx 17

INFFUS 338 No. of Pages 17, Model 5G

16 February 2009 Disk UsedARTICLE IN PRESS

umn (column 2). In other words, maximizing Z0 on row 1 can be re-duced to merely focusing on increasing the ‘‘constant” at row 1 asmuch as possible.

Therefore, the objective is to increase the ‘‘constant” at row 1 asmuch as possible. Evidently the limiting increase is given by dividingthe element in the right-hand column (referred to as the pivot ele-ment) into the element in the constant column (column 2) of the pi-vot element’s row. A value that is small in magnitude is mostrestrictive. The increase in the objective function for this choice of pi-vot element is then that value multiplied by the Z0-row entry of thatcolumn. We repeat this procedure on all possible right-hand col-umns (columns 3–13) to find the most appropriate pivot element.

For example, if we choose (row 3, column 5) – abbreviated by[R3 C5] – as the pivot element, this pivot will allow P2 to be in-creased by 2/|�1| = 2 (i.e., [R1 C5]/[R3 C5]), which results in a netincrease of the cost function (i.e., Z0-row) by an amount 0.002 � 2/|�1| = 0.004 (i.e., [R3 C2] [R1 C5]/[R3 C5]). The mechanism behindthe method by Kuenzi et al. is to always choose the pivot elementthat will lead to the smallest increase in the cost function amongall the possible choices of the pivot elements. In this case, the pivotelement is [R3 C5]. This is used to ‘‘swap in” Z1 and to ‘‘swap out”P2, resulting in the following tableau form:

213221332134

Pl(2

C1

ease009

C2

cite this), doi:10.1

C3

artic016/

C4

le inj.inf

C5

U

prefus.2

C6

ss a009

C7

N

s: B..01.0

C8

K. Sy01

C9

, In

C10

tegra

C11

ting

C12

intrus

C13

2135 Constant P0 P1 Z1 P3 P4 P5 P6 P7 S0 S1 S2 21362137 R1 Z0 �1.996 2 2 �2 1 2 3 2 3 �1 1 �1 2138 R2 Z0 0.005 �1 �1 �1 �1 1 21392140

R3

P2 0.002 �1 �1 �1 �1 1

2141

R4 Z2 0.003 �1 �1 �1 �1 1 2142 R5 Z3 0 �1 �1 21432144

R6

Z4 0.99 �1

2145

R7 Z5 0.998 �1 �1 1 �1 �1 �1 T 2146214721482149215021512152215321542155215621572158215921602161216221632164216521662167216821692170

CO

RR

ECThe pivot element swap process just described is repeated until

there is no more positive in the Z0-row (or row 1) for columnsC3–C13, which signals no further increase for Z0 in row 1 is possible.The answer for the optimal solution can then be read from the con-stant column (column 2) of the final tableau.

References

[1] A. Valdes, K. Skinner, Probabilistic Alert Correlation, LNCS, vol. 2212, RecentAdvances in Intrusion Detection, RAID 2001, Springer-Verlag.

[2] P. Stephenson, The Application of Intrusion Detection Systems in a ForensicEnvironment, RAID 2000, Springer-Verlag, Berlin, Heidelberg.

[3] B. Schneier, J. Kelsey, Secure audit logs to support computer forensics, ACMTransactions on Information and System Security 2 (2) (1999) 159–176. May,ISSN: 1094-9224.

[4] M. Roesch, Snort – Lightweight Intrusion Detection for Networks. <http://www.Snort.org/docs/lisapaper.txt>.

[5] T.H. Ptacek, T.N. Newsham, Insertion, Evasion, and Denial of Service: EludingNetwork Intrusion Detection, January 1988. <http://www.Snort.org/docs/idspaper/>.

2171

ion alert

ED

PR

OO

F

[6] <http://www.sdsc.edu/~peisert/research/peisert-dissertation.pdf>, 2007.[7] P. Sommer, Intrusion Detection System as Evidence, in: Proc. of the First

International Workshop on Recent Advances in Intrusion Detection (RAID),1998.

[8] S.T. King, P.M. Chen, Backtracking intrusions, ACM Transactions on ComputerSystems 23 (1) (2005) 51–76.

[9] S. Peisert, M. Bishop, S. Karin, K. Marzullo, Toward models for forensic analysis,in: Proceedings of the 2nd International Workshop on Systematic Approachesto Digital Forensic Engineering, 2007, pp. 3–15, ISBN:0-7695-2808-2.

[10] A. Goel, W. Feng, D. Maier, J. Walpole, Forensix: a robust, high-performancereconstruction system, in: Proceedings of the 25th International Conference onDistributed Computing Systems Workshops, 2005, pp 155–162.

[11] B.K. Sy, H. Chan, Data mining approach for anomaly detection: a signature-based approach, in: Proceedings of the 2005 International Conference on DataMining, Las Vegas, Nevada, June 20–23, 2005.

[12] D. Ellis, J. Aiken J.K. Attwood, S. Tenaglia, A behavioral approach to wormdetection, in: Proceedings of the ACM Workshop on Rapid Malcode, 2004.

[13] S. Axelsson, Research in intrusion–detection systems: a survey, TechnicalReport 98-17, December, 1998.

[14] J. Doyle, I. Kohane, W. Long, H. Shrobe, P. Szolovits, Event recognition beyondsignature and anomaly, in: Proceedings of 2001 IEEE Workshop onInformation Assurance and Security, Wert Point, NY, June 2001, pp. 17–23.

[15] P. Roberts, Nai goes Forensic with Infinistream, InforWorld, February 2003.<http://www.infoworld.com/article/03/02/10/HNnai_1.html>.

[16] Sanstorm Enterprise: Netintercept, February 2003. <http://www.sandstorm.com/products/netintercept/>.

[17] K. Shanmugasundaram, N. Memon, A. Savant, H. Bronnimann, ForNet: adistributed forensics network, in: Second International Workshop onMathematical Methods, Models, and Architectures for Computer NetworkSecurity, LNCS 2776, Springer 2003, ISBN:3-540-40797-9.

[18] S. Wu, U. Manber, A Fast Algorithm for Multi-Pattern Searching (May) (1994).[19] M. Norton, D. Roelker, Hi-Performance Multi-rule Inspection Engine,

Sourcefire Network Security Inc., 2004. April.[20] S. Axelsson, The base-rate fallacy and its implications for the difficulty of

intrusion detection, in: Proceedings of the ACM Conference on Computer andCommunication Security, November 1999. <http://www.raid-symposium.org/raid99/PAPERS/Axelsson.pdf>.

[21] N. Tuck, T. Sherwood, B. Calder, G. Varghese, Deterministic memory-efficientstring matching algorithms for intrusion detection, in: Proceedings of the IEEEInfocom Conference, Hong Kong, China, March 2004.

[22] <http://www.endace.com/>.[23] <http://www.nss.co.uk/grouptests/gigabitids/edition3/sourcefire/

sourcefire.htm>.[24] B.K. Sy, A. Gupta, Information-statistical Data Mining: Warehouse Integration

with Examples of Oracle Basics, Kluwer Academic Publishing, 2003/2004,ISBN:1-4020-7650-9.

[25] H.P. Kuenzi, H.G. Tzchach, C.A. Zehnder, Numerical Methods of MathematicalOptimization, Academic Press, New York, 1971.

[26] M.T. Health, Scientific Computing: An Introduction Survey, McGraw Hill, 1997,ISBN:0-07-027684-6.

[27] <http://www.qcwireless.net/dids_paper/svd_mechanism_illustration.pdf>.[28] W.H. Press, B.P. Flannery, S.A. Teukolsky, W.T. Vetterling, Numerical Recipes in

C: The Art of Scientific Computing, second edition, Cambridge University Press,1992, ISBN:0-521-43108-5.

[29] B.K. Sy, Probability model selection using information-theoretic optimizationcriterion, The Journal of Statistical Computation and Simulation 69(3) (2001)203–224 (also US Patent 6,961,685).

[30] <http://www.Snort.org/rules/>.[31] <http://www.bleedingSnort.com/bleeding-all.rules>.[32] <http://www.qcwireless.net/dids_paper/online_appendix.pdf>.[33] B.K. Sy, N. Mullodzhanov, Extracting forensic explanation from intrusion

alerts, in: Proc. of 2006 International Conference on Data Mining (DMIN’06),Las Vegas, Nevada, June 2006.

[34] H. Debar, M. Dacier, A. Wepsi, A Revised Taxonomy for Intrusion–DetectionSystems, IBM Research Report, 1999.

[35] J. Han, M. Kamber, Data Mining – Concepts and Techniques, Morgan KaufmannPublishers, 2001.

[36] R. Duda, P. Hart, D. Stock, Pattern Classification, John Wiley & Sons, 2001.

information to aid forensic explanation: An ..., Informat. Fusion