An Overview of Intrusion Detection & Countermeasure – Research Directions Systems

Download Report

Transcript An Overview of Intrusion Detection & Countermeasure – Research Directions Systems

An Overview of Intrusion
Detection & Countermeasure
Systems – Research Directions
Part II
Fernando C. Colon Osorio
Computer Science Department
Worcester, MA 01609
PEDS II - 10072002
1
Outline
• Previous Talk – what did we cover last?
–
–
–
–
Definitions
A Model of an Intrusion
Basic Approaches
Critical research Problems
• PEDS – Part II
– S.A.F.E Architecture
– S.A.F.E. approach to critical research problems
– Other Related topics
• Conclusions
PEDS II - 10072002
2
Intrusion Detection System –
Definition
Formal Definition [10], [11]
“Intrusion Detection (ID) is the problem of identifying individuals
who are using, or attempting to use a computer system
without authorization (i.e., crackers) and those who have
legitimate access to the system but are abusing their privileges
(i.e., the insider threat”).
PEDS II - 10072002
3
Intrusion Timeline
System is Secure/Dependable
Mth
System is Secure/Dependable


Nth
MTBASI
MTTID
MTTCI
MTBSI
Attacks Begin
System is ~ Secure/Dependable
1st
Intrusion
Attempt
Nth
2nd
Intrusion
Attempt
Intrusion
Attempt
(Success)
Intrusion
Detected by
IDS and/or
IDCS
Intrusion
Countermeasures
Launched
PEDS II - 10072002
Mth
Intrusion
Attempt
(Success)
4
Intrusion Timeline
System is Secure/Dependable
Mth
System is Secure/Dependable


Nth
Attack Is Successful
MTBASI
MTTID
MTTCI
MTBSI
System is ~ Secure/Dependable
1st
Intrusion
Attempt
Nth
2nd
Intrusion
Attempt
Intrusion
Attempt
(Success)
Intrusion
Detected by
IDS and/or
IDCS
Intrusion
Countermeasures
Launched
PEDS II - 10072002
Mth
Intrusion
Attempt
(Success)
5
Intrusion Timeline
System is Secure/Dependable
Mth
System is Secure/Dependable


Nth
Diagnosis
Region
MTBASI
MTTID
MTTCI
MTBSI
System is ~ Secure/Dependable
1st
Intrusion
Attempt
Nth
2nd
Intrusion
Attempt
Intrusion
Attempt
(Success)
Intrusion
Detected by
IDS and/or
IDCS
Intrusion
Countermeasures
Launched
PEDS II - 10072002
Mth
Intrusion
Attempt
(Success)
6
Intrusion Timeline
System is Secure/Dependable
Mth
System is Secure/Dependable


Nth
Repair/
ReIntegration
Region
MTBASI
MTTID
MTTCI
MTBSI
System is ~ Secure/Dependable
1st
Intrusion
Attempt
Nth
2nd
Intrusion
Attempt
Intrusion
Attempt
(Success)
Intrusion
Detected by
IDS and/or
IDCS
Intrusion
Countermeasures
Launched
PEDS II - 10072002
Mth
Intrusion
Attempt
(Success)
7
Intrusion Timeline
System is Secure/Dependable
Mth
System is Secure/Dependable


Nth
MTBASI
System
Operational
MTTID
MTTCI
MTBSI
System is ~ Secure/Dependable
1st
Intrusion
Attempt
Nth
2nd
Intrusion
Attempt
Intrusion
Attempt
(Success)
Intrusion
Detected by
IDS and/or
IDCS
Intrusion
Countermeasures
Launched
PEDS II - 10072002
Mth
Intrusion
Attempt
(Success)
8
Anomaly vs. Misuse IDS
systems
In past years, multiple Intrusion Detection systems
have been proposed an implemented. All of the
proposed systems are based on one or the other of
two basic approaches.
• anomaly detection
• misuse detection.
Note: Kumar [13] presents a fairly complete
categorization of the most important systems
proposed or build thus far.
PEDS II - 10072002
9
Intrusion Timeline
System is Secure/Dependable


System is Secure/Dependable
Mth
Nth
Realm of Anomalous
Detection
Techniques
MTBASI
MTTID
MTTCI
MTBSI
Realm of Misuse
Detection
Techniques
1st
Intrusion
Attempt
System is ~ Secure/Dependable
Nth
2nd
Intrusion
Attempt
Intrusion
Attempt
(Success)
Intrusion
Detected by
IDS and/or
IDCS
Intrusion
Countermeasures
Launched
PEDS II - 10072002
Mth
Intrusion
Attempt
(Success)
10
Figure 1 – Generic Intrusion Detection Model [Denning]
Audit Trails/
Network Packets/
Application Trails
S = { s1, s2, …, sn }
Assert New Rules
Modify Existing Rules
Event Generator
Environment
Update
Profile
Activity Profile
Rule Set
Generate
Anomaly
Records
Generate New Profile
Dynamically
PEDS II - 10072002
Clock
11
Problems with Current Approaches
•
Amongst the most important consideration and limitations present in the
design of all such systems are the following set of problems.
•
Problem # 1: Feature selection and pattern categorization.
–
•
Simply stated, in Denning’s Model, Figure 1, it is assumed that the event
generator can effectively select, a priori, the set of features or measures to
monitor which will render an optimal set for Intrusion Detection.
Problem # 2: the problem of adaptation.
–
Systems have been build and deployed that deal very effectively with threats or
intrusions previously reported or categorized.
–
When previously unseen threats appear, the systems perform poorly.
•
In the 1999 DARPA - Off-Line Intrusion Detection Evaluation [14], it was reported that the
systems under test failed to detect an attack in 17.2 %
PEDS II - 10072002
12
Problems, contn..
•
Problem # 3: Fault Tolerance
– Resistance to subversion: Systems do fail due to accidental or
malicious activities.
• system being designed must be able to recover from the traditional forms of
failures such as crashes, software failures, and so forth.
• System must be able to protect itself from deliberate attempts to
compromise it.
•
Problem # 4: Performance
– System must impose minimal overhead on the system is protecting
while running.
– System must be capable to sustain its performance characteristics
under increasing loads and changes in the pattern of usage.
PEDS II - 10072002
13
Problems, contn..
•
Problem # 5: Intrusion Detection System Evaluation &
Characterization
– Workloads
– Definition of “Goodness” – i.e., how reliable in detecting intrusions is
the system?
• MTBIS
• MTTR after an intrusion?
• Others?
PEDS II - 10072002
14
Secure Architecture & Fail-safe Engine
(S.A.F.E.)
S.A.F.E Intrusion Detection & Countermeasure system was
conceived with the specific goal of attacking the above
problems, and some others.
S.A.F.E is a distributed rather than a network or host based system
• Its structure is similar to the structure proposed in AAFID [ see
Balasubrayan and Garcia-Fernandez, 1998] in the sense that it
depends on a set of autonomous objects (agents in AAFID
nomenclature) that can reside anywhere in a network.
• However, S.A.F.E. differs significantly from AAFID, or other distributed
Intrusion Detection systems (see, EMERALD system), in that all control
is distributed, and implemented using a system wide distributed
system component, called the Trust Manager (TTM),
PEDS II - 10072002
15
Secure Architecture & Fail-safe Engine
(S.A.F.E.)
S.A.F.E Intrusion Detection & Countermeasure system architecture:
– Real time
– Distributed
• Most existing IDS are either host or network based. Single point of failure, lack of
scalability, poor performance
• S.A.F.E architecturally/logically partitions the functions of an IDS system into a set
of objects that reside anywhere in the system or the network.
– Hierarchical
• a set of reliable services is provided reliably and securely to the upper layers of
the architecture
– Fault-Resilient
• Internal failures
• External attacks,
PEDS II - 10072002
16
Secure Architecture & Fail-safe Engine
(S.A.F.E.)
S.A.F.E Architectural Components:
• Probes := data collection objects. Probes are started/stopped by the Event
Generator Object (EGO).
• EGO:= independent running entity that provides some of the basic S.A.F.E.
services.
– Starts/Stop probes
– Filters data from Probes. It provides data collection and a data abstraction layer (for
common architecture)
– Implements lower/level Intrusion detection functions
– Provides a Class of Event Service
•
Secure Host Manager (SHM):= independent running entity that provides two basic services
– Intrusion detection for the host
– Implements “learning” functions
– In addition, the SHM:
»
»
»
»
Provides the Trust Manager with a “Last Gasp” message – “I am Potentially Compromised” before launching
its countermeasure functions.
It stops processing further requests from the Trust Manager – takes itself off-line.
Countermeasure launching.
Starts/Stops all EGO objects
PEDS II - 10072002
17
Secure Architecture & Fail-safe Engine
(S.A.F.E.)
S.A.F.E System Wide Services:
• TTM:= the trust manager. The TTM is an independent entity
serving three major functions:
–
It “knows” which logical “nodes” can be trusted.
•
•
•
This is accomplished through the concept of a trust relationship matrix.
Nodes trust measure tij is added, modified, and changed using a distributed software algorithm
Algorithm is based on the solution to the “Byzantine General’s Problem”.
–
It delivers the “Last Gasp Message” to other nodes in the system.
–
Prevents partition of the trust system.
• Atomicity Manager (ATOM) := independent entity providing
atomic secure operations across the system.
–
running entity that provide some of the basic S.A.F.E. services.
PEDS II - 10072002
18
Secure Architecture & Fail-safe Engine (S.A.F.E.)
P
EGO
T
T
M
P
EGO
P
T
T
M
P
SHM
P
P
SHM
P
EGO
P
EGO
P
T
T
M
P
P
SHM
P
P
PEDS II - 10072002
19
S.A.F.E. SHM Intrusion Detection & Learning
Engine
G (wi )
Lower Lever Intrusion
Detection (EGO)
wi * Ci (x)
C1(x)
G (wi )
wi2* C2 (x)
( class(x), x )
C2(x)

S w * C (x)
i
.
.
.
i
True/False
(Intrusion Correctly
Identified)
G (wi )
Ck(x)
wk * Ck (x)
PEDS II - 10072002
20
Intrusion Detection Models
Figure 3 – Model of An Intrusion/ Attack
Source of Attack
Tab,
Node a
Node b
Tbd
Node c
Tae, Tea
Teb
Tce, Tec
Node e
Node d
Tef
Teg
Node f
Tdg
Tfh
Node g
Thg
Node h
Node Under Attack
PEDS II - 10072002
21
A network Model
•
A trust function Tij (t) for i ¹ j, exist between two nodes, it is not necessarily symmetrical.
•
The trust function Tij (t) changes over time.
•
In addition, the lack of trust between two nodes will be denoted as having a trust relationship of zero value, Tij (t) = 0.
•
In the above example, Node a is the source of the intruder attack, while Node h is the target of the attack. Note that, the
path for the intruder is
–
–
–
Path 1: a Ü e Ü g Ü h
Path 2: a Ü b Ü e Ü g Ü h
Path 3: a Ü d Ü g Ü h
•
This topological constraint amongst nodes in a network has a significant advantage over other approaches. That is, it
allows the designer of the IDC System to create multiple logical layers of defense against intruders, in effect, creating
time to detect potential intrusions and dwarfed them.
•
Example
–
Let’s say that nodes b and e suspect an intrusion by using traditional audit methods. Then, nodes b and e can invoke a state
change on their trust relationships with other nodes in such a way that,
Taj (t) = 0 for all j ¹ a and t > t of intrusion; and
Equation 1:
Tej (t) = 0 for all j ¹ e and t > t of intrusion.
PEDS II - 10072002
22
Problems & Well Known Solutions
Present in the IDCS field
• Problem # 1: Feature selection and pattern
categorization.
– Simply stated, in Denning’s Model, Figure 1, it is assumed
that the event generator can effectively select, a priori, the
set of features or measures to monitor which will render an
optimal set for Intrusion Detection.
PEDS II - 10072002
23
Figure 1 – Generic Intrusion Detection Model [Denning]
Audit Trails/
Network Packets/
Application Trails
S = { s1, s2, …, sn }
Assert New Rules
Modify Existing Rules
Event Generator
Environment
Update
Profile
Activity Profile
Rule Set
Generate
Anomaly
Records
Generate New Profile
Dynamically
PEDS II - 10072002
Clock
24
Learning, Feature Selection, Inductive Learning,
Learning with a teacher, KDD systems
Given a training set characterized by training samples,
T = { ( x1, y1), ( x2, y2), …, ( xn, yn) }
For some unknown function
ƒ(x) = y
Where each xi is an attribute vector of the form
Xi = { xi1, xi2, …, xik },
And each yi is the class label belonging to the set
Y = { y1, y2, …, ym }, then
Find the mapping (function), ƒ*, such that
ƒ*(x) ƒ(x)
PEDS II - 10072002
25
Learning, Feature Selection, Inductive Learning,
Learning with a teacher, KDD systems, contn…
Three Important and critical questions:
Q1: Does an ƒ*(x) exists?, i.e., is there a pattern?, is there knowledge to be mined?,
is there learning to be developed?;
Q2: What are the correct set of features xik providing “best” inference rules?
Q3: If ƒ*(x) exists, then what is the computational complexity of the algorithms trying
to find ƒ*(x)?
In this context, we are interested in algorithms that can find ƒ*(x) in finite time.
Or what is the computational quality, Q c( s), of the algorithm?
In this context, given two algorithm that find ƒ*(x), A1 and A2, then
Q c( A1) > Q c( A2) if the computational running time of A1 < A2.
PEDS II - 10072002
26
Problems & Well Known Solutions
Present in the IDCS field
• Problem # 2: the problem of adaptation.
– Systems have been build and deployed that deal very
effectively with threats or intrusions previously reported
or categorized.
– When previously unseen threats appear, the systems
perform poorly.
• In the 1999 DARPA - Off-Line Intrusion Detection
Evaluation [14], it was reported that the systems under test
failed to detect an attack in 17.2 %
PEDS II - 10072002
27
Figure 1 – Generic Intrusion Detection Model [Denning]
Audit Trails/
Network Packets/
Application Trails
S = { s1, s2, …, sn }
Assert New Rules
Modify Existing Rules
Event Generator
Environment
Update
Profile
Activity Profile
Rule Set
Generate
Anomaly
Records
Generate New Profile
Dynamically
PEDS II - 10072002
Clock
28
Figure 2 – A simplified Intrusion Detection Engine
S = { s1, s2, …, sn }
Decision Engine
fg (y, S, M, P(n), T, G )
Environment
Clock
a = {a1, a2, …, an }
Memory of IDS
(Rule Set/ Activity Profile
Create New Rules/Profiles
Modify Existing Rules/Profiles
PEDS II - 10072002
29
Meta - Learning
Loose Definition [chang’ 93]:
“Learning from learned knowledge”
References:
[chang’ 93]:= P. Chang and S. Stolfo. “Experiments on multistrategy learning by meta-learning”, Proc. Second
International Workshop, Multistrategy Learning, pp. 150-165, 1993.
PEDS II - 10072002
30
Meta – Learning, contn…
Let the training set be instances of correct
classifications and predictions, such as the
Training set is of the form:
T = { class( x), C1(x), C2(x), …, Ck(x) | x E }
PEDS II - 10072002
31
S.A.F.E. SHM Intrusion Detection & Learning
Engine
G (wi )
wi * Ci (x)
C1(x)
G (wi )
wi2* C2 (x)
( class(x), x )
C2(x)

S w * C (x)
i
.
.
.
i
True/False
(Intrusion Correctly
Identified)
G (wi )
Ck(x)
wk * Ck (x)
PEDS II - 10072002
32
Problems, contn..
•
Problem # 3: Fault Tolerance
– Resistance to subversion: Systems do fail due to accidental or
malicious activities.
• The Trust Manager (TTM)
•
Problem # 4: Performance
– Distributed architecture
– Light weight processes
– Simple objects
PEDS II - 10072002
33
Problems, contn..
•
Problem # 5: Intrusion Detection Systems Evaluation &
Characterization
– Workloads – testing, e.g., the 1999 DARPA evaluation uses 10
workload or traces to compare systems
• Results are empirical
• Not representative of all environments
• Very little data available
– Definition of “Goodness” (Modeling)
• how reliable in detecting intrusions is the system?, and
• how intrusion resilient is the underlying system being protected?
PEDS II - 10072002
34
Intrusion Timeline
System

 is Secure/Dependable
Nth
System is Secure/Dependable
Mth
MTTR
MTBASI
MTTID
MTTCI
MTBSI
Attacks Begin
System is ~ Secure/Dependable
1st
Intrusion
Attempt
Nth
2nd
Intrusion
Attempt
Intrusion
Attempt
(Success)
Intrusion
Detected by
IDS and/or
IDCS
Intrusion
Countermeasures
Launched
PEDS II - 10072002
Mth
Intrusion
Attempt
(Success)
35
Problem # 5: Intrusion Detection Systems Evaluation &
Characterization, contn..
•
Definition of “Goodness” – i.e., how reliable is the system with
respect to intrusions?
• MTBFi – mean time between failures/intrusion (defn: in this context a failure
is a successful intrusion) – a measure of the underlying system being
protected.
• MTTID – mean time to intrusion detection
• MTTCI – mean time to countermeasure issuance
• MTTR – mean time to contain & repair a successful intrusion
• Of course – Availability or A(t) = P ({system is operational and intrusion free
at time t1 if it was intrusion free at time t0 }
Other metrics:
• MTBASI – mean time between 1st attack and successful intrusion
PEDS II - 10072002
36
Problem # 5: Intrusion Detection Systems Evaluation &
Characterization, contn..
•
Definition of “Goodness” – i.e., how reliable is the system with respect
to intrusions?
• MTBFi – mean time between failures/intrusion (defn: in this context a failure is a
successful intrusion)
– A measure of the quality of the software system (O/S, Applications, and so forth)
• Further assume that MTBFi a
Reliability of the software, then,
– Techniques such as Component Base Reliability Estimation (CBRE), see Krishnamurthy
and Mathur;
– Software Test Coverage and Reliability techniques, see Malaiya and Karcich- where
software testing & coverage is used as
» predictors of software reliability
» To estimate the remaining defects or number of residual faults; and
» Mean time between failures or bugs.
Are all applicable.
• Challenge – to define an accurate model!!!
PEDS II - 10072002
37
Other related Topics – Honeypots &
Honeynets
Honeypot:
•
A honeypot is a fake or false system to lure the hacker into. It provides
another obstacle for the hacker.
•
honeypot systems are decoy servers or systems set up to gather
information regarding an attacker or intruder into your system.
•
honeypot traps tempt intruders into areas which appear attractive, worth
investigating and easy to access, taking them away from the really
sensitive areas of your systems. They do not replace other traditional
Internet security systems but act as an additional safeguard with alarms.
•
A honeypot is a resource which pretends to be a real target. A honeypot is
expected to be attacked or compromised. The main goals are the
distraction of an attacker and the gain of information about an attack and
the attacker.
PEDS II - 10072002
38
honeypots
honeypots will help you:
• notice when you are penetrated
• learn how attacks are formed
• identify who is attacking you
PEDS II - 10072002
39
honeypot Examples
• honeypot Project
– http://www.landfield.com/isn/mail-archive/2000/Nov/0124.html
• Deception Tool Kit Project
– http://www.all.net/dtk/index.html
• Specter
– http://www.specter.com/default50.htm
PEDS II - 10072002
40
“Specter” – Basic Idea
• Virtual Machine (VM) environment
– Early Traps
– Early detection
PEDS II - 10072002
41
honeypot Tools – “Specter”
PEDS II - 10072002
42
Honeypots Limitations
• Hard to Maintain
• Human Resource Intensive – Specialize
Knowledge
– Operating Systems
– Network security
– Current deficiencies (holes) in both O/S
and applications
PEDS II - 10072002
43
Honeynet
Honeynet Honeypots
Honeynet (Defn)
• A network system
• All systems are standard production
systems
• All usage is ~ Production
PEDS II - 10072002
44
Honeynet
PEDS II - 10072002
45
Conclusions
• A new model based on Byzantine
General’s problem will be
investigated.
• Research Area is prime for
discovery.
PEDS II - 10072002
46