1 / 28

Internet Traffic Filtering System based on Data Mining Approach

Internet Traffic Filtering System based on Data Mining Approach. Vladimir Maslyakov Computer Science Department of Lomonosov Moscow State University. Plan. Actuality Current Approaches Our Approach Results. Actuality.

marin
Download Presentation

Internet Traffic Filtering System based on Data Mining Approach

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Internet Traffic Filtering System based on Data Mining Approach Vladimir Maslyakov Computer Science Department of Lomonosov Moscow State University

  2. Plan • Actuality • Current Approaches • Our Approach • Results

  3. Actuality • About 30% of its work time knowledge workers spend on personal usage: surfing Internet, chatting, entertaining, etc* • Only 40% of organizations use some software to control Internet traffic • 90% of organizations would like to use some sort of software to control Internet traffic** * Dan Malachowski (2005, July 11) Salary.com “Wasting Time at Work Costing Companies Billions” ** Karl Donert, Sara Carro Martinez (2002, December 23) “End-user Requirements: Final Report”

  4. Goals • to prevent access to unwanted content. A common scenario for schools and libraries; • to detect harmful content (viruses, spam, trojans, etc) or links to such content; • to detect possible intrusion threats or suspicious traffic; • to reduce number of leakages of confidential information; • to prevent unwanted usage of Internet during working time.

  5. Problems Modern Internet: • is HUGE • is constantly changing (in size and content) • mostly consists of dynamic resources Organizations want: • to be independent of knowledge bases from third-party companies • a good precision of filtering with minimum number of errors • maximum performance with minimum resources

  6. Requirements • the ability to process and filter information online, which means that end users should not experience significant delays because of traffic filtering; • high precision of filtering (low rates of false-positive and false-negative errors); • the ability to process dynamic content; • the ability to adapt to new types of resources and filter resources using its content as well as its metadata; • scalability, the ability to be installed in organizations of different scale.

  7. Desirable features • Independence of language of Internet resources • Configurability and ability to adapt to specific organization needs • Ability to analyze traffic transparently to end users

  8. Plan • Actuality • Current Approaches • Our Approach • Results

  9. Traffic Filtering Systems Classification

  10. Characteristics • speed of filtering (kbytes/sec); • false-positive errors, cases when system forbids access to legal resource (%); • false-negative errors, cases when system allows access to forbidden resource (%); • precision of filtering, ratio of correctly allowed and correctly forbidden resources to the total number of analyzed resources (%).

  11. Problems • There is no benchmarks and methodologies for testing Internet Traffic Filtering Software • There is even no datasets on which tests could be performed

  12. Signature Approach • Bring signature database of analyzed resources to up-to-date state. • Process requests from users in real-time. • If Internet resource A is marked as harmful, access to the resource is forbidden. But Content at the update stage and content at the stage of analysis could be different!

  13. Advantages of Signature Approach • The signature database is usually centralized and is maintained by some organization • Content analysis is often not used at all => very good speed • Human experts precision is very good (>95%)

  14. Disadvantages of Signature Approach • speed and precision of analysis significantly depends on quality of centralized knowledge base • analysis of Internet resources without using its content or poor precision of content analysis • inability to add new types and categories of resources

  15. Other drawbacks of modern solutions • inability to personalize traffic • inability to analyze resources in different languages • inability to analyze dynamic content online • outdated databases • etc

  16. Plan • Actuality • Current Approaches • Our Approach • Results

  17. Architecture

  18. Components of System • Cache Proxy Server (Squid with ICAP support) • Java-based Kernel with built-in ICAP client • Java-based Decision Making Module • C++-based Parser (antlr) • C++-based Classifier (Perceptron, SVM) • XML-RPC for interoperability between different components (apache xml-rpc)

  19. Data Mining Approach • A fully automatic process of training on test data. • All requests are identified and transparently redirected from cache proxy server to the kernel of the system. • Kernel saves all information about the requested resource and the request’s author. • Parser tokenizes html content and extracts all the links • Classifier assigns categories to requested resource, based on output of parser, and returns results to Decision Making Module • Decision Making Module decides whether to allow or forbid current resource to user, basing on classification labels, resource metadata and user rights.

  20. Additional features • Analysis of resource in offline using robot to extract information about hyperlinks => higher precision • Ability to make incrementive learning of algorithm • Ability to parse using n-gramm methods => independence of language (payback - lower precision) • User identification by ip address

  21. Advantages of Data Mining Approach • necessary speed of analysis (decimal fractions of a second); • necessary precision of analysis (more than 90% with 1% of false-positive errors); • adaptation and self-learning, which will allow to adapt to specific organization needs; • scalability of result system, which will allow to deploy system in organizations of different scale; • autonomy from external knowledge bases and human experts.

  22. Disadvantages of Data Mining Approach • higher rate of false-positive errors than signature approach • necessity of training data

  23. Future Plans • to develop rdf knowledge base of users, statistics and resources (based on HP Jena or Protégé) • to enhance identification algorithms (SSO and LDAP support) • to enhance decision making strategies (bayesian networks or svm) • to implement more sophisticated user rights system

  24. Plan • Actuality • Current Approaches • Our Approach • Results

  25. What we have • A developed prototype • Pentium III 550 Mhz with 384 Mbs of RAM • Debian 4.0 Etch • Java 1.5.0 • Tomcat 5.0 • Gcc 4.1.2 • Some other third-party libraries (antlr 2.7.7, apache xml-rpc 3.0, berkeley db 4.5.20, etc)

  26. Training Corpus • Bank Research Corpus • 11 themes (Java, Visual Basic, Astronomy, …) • 11000 documents, 1000 for each theme

  27. Results • Training average speed: 1.3 seconds per document – ln(n) • Average Download Speed: 2 seconds • Parsing & Classification average speed: 0.6 seconds – n • Accuracy: ~90% • False-positive errors: ~1-2%

  28. Thank You!Q & A

More Related