Current Search: Computer Science (x)
View All Items
Pages
- Title
- Department of Computer & Electrical Engineering and Computer Science: The entrepreneurial department.
- Creator
- Furht, Borko
- PURL
- http://purl.flvc.org/fcla/dt/332873
- Subject Headings
- Computer science--Research.
- Format
- Document (PDF)
- Title
- Diffie-Hellman key exchange protocol, its generalization and nilpotent groups.
- Creator
- Mahalanobis, Ayan, Florida Atlantic University, Charles E. Schmidt College of Science, Department of Mathematical Sciences
- Abstract/Description
-
This dissertation has two chapters. In the first chapter we talk about the discrete logarithm problem, more specifically we concentrate on the Diffie-Hellman key exchange protocol. We survey the current state of security for the Diffie-Hellman key exchange protocol. We also motivate the reader to think about the Diffie-Hellman key exchange in terms of group automorphisms. In the second chapter we study two key exchange protocols similar to the Diffie-Hellman key exchange protocol using an...
Show moreThis dissertation has two chapters. In the first chapter we talk about the discrete logarithm problem, more specifically we concentrate on the Diffie-Hellman key exchange protocol. We survey the current state of security for the Diffie-Hellman key exchange protocol. We also motivate the reader to think about the Diffie-Hellman key exchange in terms of group automorphisms. In the second chapter we study two key exchange protocols similar to the Diffie-Hellman key exchange protocol using an abelian subgroup of the automorphism group of a non-abelian group. We also generalize group no. 92 of the Hall-Senior table, for arbitrary prime p and study the automorphism group of these generalized group. We show that for those groups, the group of central automorphisms is an abelian group. We use these central automorphisms for the key exchange we are studying. We also develop a signature scheme.
Show less - Date Issued
- 2005
- PURL
- http://purl.flvc.org/fcla/dt/12154
- Subject Headings
- Mathematics, Computer Science
- Format
- Document (PDF)
- Title
- Evolution of numeric constants in Genetic Programming.
- Creator
- Fernandez, Thomas, Florida Atlantic University, Evett, Matthew P.
- Abstract/Description
-
Genetic Programming is an evolutionary technique for searching through the space of S-expressions for programs that represent optimal or acceptable solutions to a given problem. Genetic Programming often has difficulty in finding the appropriate numeric constants to use in leaf nodes of the S-expressions. This thesis describes the use of local search algorithms to search for numeric constants that will improve the S-expressions found by Genetic Programming. Three methods, Multi-Dimensional...
Show moreGenetic Programming is an evolutionary technique for searching through the space of S-expressions for programs that represent optimal or acceptable solutions to a given problem. Genetic Programming often has difficulty in finding the appropriate numeric constants to use in leaf nodes of the S-expressions. This thesis describes the use of local search algorithms to search for numeric constants that will improve the S-expressions found by Genetic Programming. Three methods, Multi-Dimensional Hill Climbing, Vector Hill Climbing, and Numeric Mutation are combined with Genetic Programming to create hybrid systems. The performance of these hybrid systems is analyzed and future directions for improving Genetic Programming with the use of hybrid systems are discussed.
Show less - Date Issued
- 1997
- PURL
- http://purl.flvc.org/fcla/dt/15493
- Subject Headings
- Genetic programming (Computer science)
- Format
- Document (PDF)
- Title
- An artificial neural network architecture for interpolation, function approximation, time series modeling and control applications.
- Creator
- Luebbers, Paul Glenn., Florida Atlantic University, Pandya, Abhijit S., Sudhakar, Raghavan, College of Engineering and Computer Science, Department of Computer and Electrical Engineering and Computer Science
- Abstract/Description
-
A new artificial neural network architecture called Power Net (PWRNET) and Orthogonal Power Net (OPWRNET) has been developed. Based on the Taylor series expansion of the hyperbolic tangent function, this novel architecture can approximate multi-input multi-layer artificial networks, while requiring only a single layer of hidden nodes. This allows a compact network representation with only one layer of hidden layer weights. The resulting trained network can be expressed as a polynomial...
Show moreA new artificial neural network architecture called Power Net (PWRNET) and Orthogonal Power Net (OPWRNET) has been developed. Based on the Taylor series expansion of the hyperbolic tangent function, this novel architecture can approximate multi-input multi-layer artificial networks, while requiring only a single layer of hidden nodes. This allows a compact network representation with only one layer of hidden layer weights. The resulting trained network can be expressed as a polynomial function of the input nodes. Applications which cannot be implemented with conventional artificial neural networks, due to their intractable nature, can be developed with these network architectures. The degree of nonlinearity of the network can be directly controlled by adjusting the number of hidden layer nodes, thus avoiding problems of over-fitting which restrict generalization. The learning algorithm used for adapting the network is the familiar error back propagation training algorithm. Other learning algorithms may be applied and since only one hidden layer is to be trained, the training performance of the network is expected to be comparable to or better than conventional multi-layer feed forward networks. The new architecture is explored by applying OPWRNET to classification, function approximation and interpolation problems. These applications show that the OPWRNET has comparable performance to multi-layer perceptrons. The OPWRNET was also applied to the prediction of noisy time series and the identification of nonlinear systems. The resulting trained networks, for system identification tasks, can be expressed directly as discrete nonlinear recursive polynomials. This characteristic was exploited in the development of two new neural network based nonlinear control algorithms, the Linearized Self-Tuning Controller (LSTC) and a variation of a Neural Adaptive Controller (NAC). These control algorithms are compared to a linear self-tuning controller and an artificial neural network based Inverse Model Controller. The advantages of these new controllers are discussed.
Show less - Date Issued
- 1994
- PURL
- http://purl.flvc.org/fcla/dt/12357
- Subject Headings
- Neural networks (Computer science)
- Format
- Document (PDF)
- Title
- The effect of compression of performance in a demand paging operating system.
- Creator
- Wynn, Allen Chester., Florida Atlantic University, Wu, Jie, College of Engineering and Computer Science, Department of Computer and Electrical Engineering and Computer Science
- Abstract/Description
-
In this thesis, we measure and analyze the effects of compression in a demand paging operating system. We first explore existing compression algorithms and page replacement policies currently in use. Then we examine the OS/2 operating system which is modified to include page-based compression. Software trace hooks are inserted into the operating system to determine the amount of time required to process a page fault for each type of page, e.g. non-compressed, compressed, zero-filled, and the...
Show moreIn this thesis, we measure and analyze the effects of compression in a demand paging operating system. We first explore existing compression algorithms and page replacement policies currently in use. Then we examine the OS/2 operating system which is modified to include page-based compression. Software trace hooks are inserted into the operating system to determine the amount of time required to process a page fault for each type of page, e.g. non-compressed, compressed, zero-filled, and the number of page faults for each type of page. Software trace measurements as well as physical timings are taken on a system without compressed pages and the same system with compressed pages. We find the system with compressed pages shows a slight increase in paging activity for memory constrained systems, but performance (time) is improved in both memory constrained and unconstrained systems.
Show less - Date Issued
- 1997
- PURL
- http://purl.flvc.org/fcla/dt/15421
- Subject Headings
- Paging (Computer science), Data compression (Computer science), Operating systems (Computers)
- Format
- Document (PDF)
- Title
- Quantum-Resistant Key Agreement and Key Encapsulation.
- Creator
- Robinson, Angela, Steinwandt, Rainer, Florida Atlantic University, Charles E. Schmidt College of Science, Department of Mathematical Sciences
- Abstract/Description
-
We explore quantum-resistant key establishment and hybrid encryption. We nd that while the discrete logarithm problem is e ciently solved by a quantum computer using Shor's algorithm, some instances are insecure even using classical computers. The discrete logarithm problem based on a symmetric group Sn is e - ciently solved in polynomial time. We design a PUF-based 4-round group key establishment protocol, adjusting the model to include a physical channel capable of PUF transmission, and...
Show moreWe explore quantum-resistant key establishment and hybrid encryption. We nd that while the discrete logarithm problem is e ciently solved by a quantum computer using Shor's algorithm, some instances are insecure even using classical computers. The discrete logarithm problem based on a symmetric group Sn is e - ciently solved in polynomial time. We design a PUF-based 4-round group key establishment protocol, adjusting the model to include a physical channel capable of PUF transmission, and modify adversarial capabilities with respect to the PUFs. The result is a novel group key establishment protocol which avoids computational hardness assumptions and achieves key secrecy. We contribute a hybrid encryption scheme by combining a key encapsulation mechanism (KEM) with a symmetric key encryption scheme by using two hash functions. We require only one-way security in the quantum random oracle model (QROM) of the KEM and one-time security of the symmetric encryption scheme in the QROM. We show that this hybrid scheme is IND-CCA secure in the QROM. We rely on a powerful theorem by Unruh that provides an upper bound on indistinguishability between the output of a random oracle and a random string, when the oracle can be accessed in quantum superposition. Our result contributes to the available IND-CCA secure encryption schemes in a setting where quantum computers are under adversarial control. Finally, we develop a framework and describe biometric visual cryptographic schemes generically under our framework. We formalize several security notions and de nitions including sheet indistinguishability, perfect indistinguishability, index recovery, perfect index privacy, and perfect resistance against false authentication. We also propose new and generic strategies for attacking e-BVC schemes such as new distinguishing attack, new index recovery, and new authentication attack. Our quantitative analysis veri es the practical impact of our framework and o ers concrete upper bounds on the security of e-BVC.
Show less - Date Issued
- 2018
- PURL
- http://purl.flvc.org/fau/fd/FA00013023
- Subject Headings
- Quantum computing, Data encryption (Computer science), Cryptography
- Format
- Document (PDF)
- Title
- Complexity metrics in parallel computing.
- Creator
- Larrondo-Petrie, Maria M., Florida Atlantic University, Fernandez, Eduardo B., Coulter, Neal S., College of Engineering and Computer Science, Department of Computer and Electrical Engineering and Computer Science
- Abstract/Description
-
Accompanying the potential increase in power offered by parallel computers is an increase in the complexity of program design, implementation, testing and maintenance. It is important to understand the logical complexity of parallel programs in order to support the development of concurrent software. Measures are needed to quantify the components of parallel software complexity and to establish a basis for comparison and analysis of parallel algorithms at various stages of development and...
Show moreAccompanying the potential increase in power offered by parallel computers is an increase in the complexity of program design, implementation, testing and maintenance. It is important to understand the logical complexity of parallel programs in order to support the development of concurrent software. Measures are needed to quantify the components of parallel software complexity and to establish a basis for comparison and analysis of parallel algorithms at various stages of development and implementation. A set of primitive complexity measures is proposed that collectively describe the total complexity of parallel programs. The total complexity is separated into four dimensions or components: requirements, sequential, parallel and communication. Each proposed primitive measure is classified under one of these four areas. Two additional possible dimensions, fault-tolerance and real-time, are discussed. The total complexity measure is expressed as a vector of dimensions; each component is defined as a vector of primitive metrics. The method of quantifying each primitive metric is explained in detail. Those primitive metrics that contribute to the parallel and communications complexity are exercised against ten published summation algorithms and programs, illustrating that architecture has a significant effect on the complexity of parallel programs--even if the same programming language is used. The memory organization and the processor interconnection scheme had no effect on the parallel component, but did affect the communication component. Programming style and language did not have a noticeable effect on either component. The proposed metrics are quantifiable, consistent, and useful in comparing parallel algorithms. Unlike existing parallel metrics, they are general and applicable to different languages, architectures, algorithms, paradigms, programming styles and stages of software development.
Show less - Date Issued
- 1992
- PURL
- http://purl.flvc.org/fcla/dt/12296
- Subject Headings
- Parallel programming (Computer Science), Computer algorithms
- Format
- Document (PDF)
- Title
- IBM PC AT FILE SYSTEM PERFORMANCE EVALUATION.
- Creator
- PHILLIPS, DEBRA ANN., Florida Atlantic University, Coulter, Neal S.
- Abstract/Description
-
This thesis gives an evaluation of DOS 3.2 file system performance in the current IBM PC AT environment, and it presents a survey of alternative file system and high density storage integration strategies. The current file system is evaluated to determine the nature of its algorithms and structures. In particular, the file system is examined from a disk access perspective and from the perspective of alternative disk and file management strategies used in UNIX*2 file systems.
- Date Issued
- 1987
- PURL
- http://purl.flvc.org/fcla/dt/14379
- Subject Headings
- IBM Personal Computer, File organization (Computer science)
- Format
- Document (PDF)
- Title
- Novel Techniques in Genetic Programming.
- Creator
- Fernandez, Thomas, Furht, Borko, Florida Atlantic University, College of Engineering and Computer Science, Department of Computer and Electrical Engineering and Computer Science
- Abstract/Description
-
Three major problems make Genetic Programming unfeasible or impractical for real world problems. The first is the excessive time complexity.In nature the evolutionary process can take millions of years, a time frame that is clearly not acceptable for the solution of problems on a computer. In order to apply Genetic Programming to real world problems, it is essential that its efficiency be improved. The second is called overfitting (where results are inaccurate outside the training data). In a...
Show moreThree major problems make Genetic Programming unfeasible or impractical for real world problems. The first is the excessive time complexity.In nature the evolutionary process can take millions of years, a time frame that is clearly not acceptable for the solution of problems on a computer. In order to apply Genetic Programming to real world problems, it is essential that its efficiency be improved. The second is called overfitting (where results are inaccurate outside the training data). In a paper[36] for the Federal Reserve Bank, authors Neely and Weller state “a perennial problem with using flexible, powerful search procedures like Genetic Programming is overfitting, the finding of spurious patterns in the data. Given the well-documented tendency for the genetic program to overfit the data it is necessary to design procedures to mitigate this.” The third is the difficulty of determining optimal control parameters for the Genetic Programming process. Control parameters control the evolutionary process. They include settings such as, the size of the population and the number of generations to be run. In his book[45], Banzhaf describes this problem, “The bad news is that Genetic Programming is a young field and the effect of using various combinations of parameters is just beginning to be explored.” We address these problems by implementing and testing a number of novel techniques and improvements to the Genetic Programming process. We conduct experiments using data sets of various degrees of difficulty to demonstrate success with a high degree of statistical confidence.
Show less - Date Issued
- 2006
- PURL
- http://purl.flvc.org/fau/fd/FA00012570
- Subject Headings
- Evolutionary programming (Computer science), Genetic algorithms, Genetic programming (Computer science)
- Format
- Document (PDF)
- Title
- Generating formal models from UML class diagrams.
- Creator
- Shroff, Malcolm Keki., Florida Atlantic University, France, Robert B., Larrondo-Petrie, Maria M., College of Engineering and Computer Science, Department of Computer and Electrical Engineering and Computer Science
- Abstract/Description
-
The rich structuring mechanisms, and abstract modeling constructs available in most graphical object-oriented modeling methods (OOMs) facilitate the creation of abstract, visually-appealing, highly-structured graphical models. On the other hand, lack of formal semantics for the modeling notation can severely limit the utility of OOMs. Formal specification techniques (FSTs) support the creation of precise and analyzable specifications, but they can be tedious to create and difficult to read,...
Show moreThe rich structuring mechanisms, and abstract modeling constructs available in most graphical object-oriented modeling methods (OOMs) facilitate the creation of abstract, visually-appealing, highly-structured graphical models. On the other hand, lack of formal semantics for the modeling notation can severely limit the utility of OOMs. Formal specification techniques (FSTs) support the creation of precise and analyzable specifications, but they can be tedious to create and difficult to read, especially by system developers not trained in formal methods. The complementary strengths of OOMs and FSTs suggest that their integration can result in techniques that can be used to create precise and analyzable models. This thesis describes a technique for integrating analysis level UML (Unified Modeling Language) Class Diagrams with the formal notation Object-Z.
Show less - Date Issued
- 1997
- PURL
- http://purl.flvc.org/fcla/dt/15514
- Subject Headings
- UML (Computer science), Object-oriented methods (Computer science)
- Format
- Document (PDF)
- Title
- The notion of aggregation.
- Creator
- Saksena, Monika., Florida Atlantic University, France, Robert B., Larrondo-Petrie, Maria M., College of Engineering and Computer Science, Department of Computer and Electrical Engineering and Computer Science
- Abstract/Description
-
Most popular object-oriented modeling techniques (OOMTs) provide good support for the creation of conceptual models of system behavior and structure. A serious drawback of these techniques is that the concepts and notations used are not rigorously defined. This can lead to the creation of ambiguous models, and to disagreements over the proper use and interpretation of modeling constructs. An important modeling construct that is often loosely defined is aggregation. This thesis presents a...
Show moreMost popular object-oriented modeling techniques (OOMTs) provide good support for the creation of conceptual models of system behavior and structure. A serious drawback of these techniques is that the concepts and notations used are not rigorously defined. This can lead to the creation of ambiguous models, and to disagreements over the proper use and interpretation of modeling constructs. An important modeling construct that is often loosely defined is aggregation. This thesis presents a precise characterization of aggregation that can help developers identify appropriate applications of the concept. Our characterization is the result of careful analysis of literature on conceptual modeling, knowledge representation and object-oriented (OO) modeling. We discuss primary and secondary properties of aggregation and propose annotations for UML (Unified Modeling Language). An extensive discussion of the more useful patterns of aggregation helps developers pick a suitable prescription of aggregation.
Show less - Date Issued
- 1998
- PURL
- http://purl.flvc.org/fcla/dt/15537
- Subject Headings
- Object-oriented methods (Computer science), UML (Computer science)
- Format
- Document (PDF)
- Title
- The magic of hypercube: from parallel computing to multimedia.
- Creator
- Furht, Borko
- Date Issued
- 2009-07
- PURL
- http://purl.flvc.org/fcla/dt/332893
- Subject Headings
- Computer science --Research., Digital video.
- Format
- Document (PDF)
- Title
- Power based wide collision attacks on AES.
- Creator
- Ye, Xin, Eisenbarth, Thomas, Graduate College
- Date Issued
- 2011-04-08
- PURL
- http://purl.flvc.org/fcla/dt/3164806
- Subject Headings
- Computer networks, Data encryption (Computer science), Computer security
- Format
- Document (PDF)
- Title
- A BCU scalable sensory acquisition system for EEG embedded applications.
- Creator
- Fathalla, Sherif S., College of Engineering and Computer Science, Department of Computer and Electrical Engineering and Computer Science
- Abstract/Description
-
Electroencephalogram (EEG) Recording has been through a lot of changes and modification since it was first introduced in 1929 due to rising technologies and signal processing advancements. The EEG Data acquisition stage is the first and most valuable component in any EEG recording System, it has the role of gathering and conditioning its input and outputting reliable data to be effectively analyzed and studied by digital signal processors using sophisticated and advanced algorithms which help...
Show moreElectroencephalogram (EEG) Recording has been through a lot of changes and modification since it was first introduced in 1929 due to rising technologies and signal processing advancements. The EEG Data acquisition stage is the first and most valuable component in any EEG recording System, it has the role of gathering and conditioning its input and outputting reliable data to be effectively analyzed and studied by digital signal processors using sophisticated and advanced algorithms which help in numerous medical and consumer applications. We have designed a low noise low power EEG data acquisition system that can be set to act as a standalone mobile EEG data processing unit providing data preprocessing functions; it can also be a very reliable high speed data acquisition interface to an EEG processing unit.
Show less - Date Issued
- 2010
- PURL
- http://purl.flvc.org/FAU/3164095
- Subject Headings
- Brain-computer interfaces, Computational neuroscience, Neural networks (Computer science)
- Format
- Document (PDF)
- Title
- Quantum Circuits for Symmetric Cryptanalysis.
- Creator
- Langenberg, Brandon Wade, Steinwandt, Rainer, Florida Atlantic University, Charles E. Schmidt College of Science, Department of Mathematical Sciences
- Abstract/Description
-
Quantum computers and quantum computing is a reality of the near feature. Companies such as Google and IBM have already declared they have built a quantum computer and tend to increase their size and capacity moving forward. Quantum computers have the ability to be exponentially more powerful than classical computers today. With this power modeling behavior of atoms or chemical reactions in unusual conditions, improving weather forecasts and traffic conditions become possible. Also, their...
Show moreQuantum computers and quantum computing is a reality of the near feature. Companies such as Google and IBM have already declared they have built a quantum computer and tend to increase their size and capacity moving forward. Quantum computers have the ability to be exponentially more powerful than classical computers today. With this power modeling behavior of atoms or chemical reactions in unusual conditions, improving weather forecasts and traffic conditions become possible. Also, their ability to exponentially speed up some computations makes the security of todays data and items a major concern and interest. In the area of cryptography, some encryption schemes (such as RSA) are already deemed broken by the onset of quantum computing. Some encryption algorithms have already been created to be quantum secure and still more are being created each day. While these algorithms in use today are considered quantum-safe not much is known of what a quantum attack would look like on these algorithms. Specifically, this paper discusses how many quantum bits, quantum gates and even the depth of these gates that would be needed for such an attack. The research below was completed to shed light on these areas and offer some concrete numbers of such an attack.
Show less - Date Issued
- 2018
- PURL
- http://purl.flvc.org/fau/fd/FA00013010
- Subject Headings
- Quantum computing, Cryptography, Cryptanalysis, Data encryption (Computer science), Computer algorithms
- Format
- Document (PDF)
- Title
- Learning in connectionist networks using the Alopex algorithm.
- Creator
- Venugopal, Kootala Pattath., Florida Atlantic University, Pandya, Abhijit S., Sudhakar, Raghavan, College of Engineering and Computer Science, Department of Computer and Electrical Engineering and Computer Science
- Abstract/Description
-
The Alopex algorithm is presented as a universal learning algorithm for connectionist models. It is shown that the Alopex procedure could be used efficiently as a supervised learning algorithm for such models. The algorithm is demonstrated successfully on a variety of network architectures. Such architectures include multilayer perceptrons, time-delay models, asymmetric, fully recurrent networks and memory neuron networks. The learning performance as well as the generation capability of the...
Show moreThe Alopex algorithm is presented as a universal learning algorithm for connectionist models. It is shown that the Alopex procedure could be used efficiently as a supervised learning algorithm for such models. The algorithm is demonstrated successfully on a variety of network architectures. Such architectures include multilayer perceptrons, time-delay models, asymmetric, fully recurrent networks and memory neuron networks. The learning performance as well as the generation capability of the Alopex algorithm are compared with those of the backpropagation procedure, concerning a number of benchmark problems, and it is shown that the Alopex has specific advantages over the backpropagation. Two new architectures (gain layer schemes) are proposed for the on-line, direct adaptive control of dynamical systems using neural networks. The proposed schemes are shown to provide better dynamic response and tracking characteristics, than the other existing direct control schemes. A velocity reference scheme is introduced to improve the dynamic response of on-line learning controllers. The proposed learning algorithm and architectures are studied on three practical problems; (i) Classification of handwritten digits using Fourier Descriptors; (ii) Recognition of underwater targets from sonar returns, considering temporal dependencies of consecutive returns and (iii) On-line learning control of autonomous underwater vehicles, starting with random initial conditions. Detailed studies are conducted on the learning control applications. Effect of the network learning rate on the tracking performance and dynamic response of the system are investigated. Also, the ability of the neural network controllers to adapt to slow and sudden varying parameter disturbances and measurement noise is studied in detail.
Show less - Date Issued
- 1993
- PURL
- http://purl.flvc.org/fcla/dt/12325
- Subject Headings
- Computer algorithms, Computer networks, Neural networks (Computer science), Machine learning
- Format
- Document (PDF)
- Title
- Comparison of embedded security versus separated security in an application.
- Creator
- Oviedo, Carlos., Florida Atlantic University, VanHilst, Michael
- Abstract/Description
-
This work discusses and compares two different approaches that design and implement a requirement for security in an application. The construction process followed for the security features determines how easily further changes can be accommodated, after the application has been built. How the problem is decomposed into modules, and when, determines if a solution or parts from the solution will be reusable without modification in the same application after changes have been made to address a...
Show moreThis work discusses and compares two different approaches that design and implement a requirement for security in an application. The construction process followed for the security features determines how easily further changes can be accommodated, after the application has been built. How the problem is decomposed into modules, and when, determines if a solution or parts from the solution will be reusable without modification in the same application after changes have been made to address a new or altered requirement. Two construction perspectives are analyzed. In the first perspective, security features are embedded within the application design. In the second approach, the security design is separated from the rest of the application. For this latter implementation, an aspect oriented approach is used. The analysis performed shows that how the problem is decomposed leads to different designs, which present different levels of challenge for the application's future evolution. If a more adaptable solution can be designed and implemented, then the application will be more flexible to accommodate new changes and, as a consequence, be more reusable.
Show less - Date Issued
- 2006
- PURL
- http://purl.flvc.org/fcla/dt/13324
- Subject Headings
- Data encryption (Computer science), Computer security, Embedded computer systems
- Format
- Document (PDF)
- Title
- Modeling and analysis of security.
- Creator
- Ajaj, Ola, Fernandez, Eduardo B., College of Engineering and Computer Science, Department of Computer and Electrical Engineering and Computer Science
- Abstract/Description
-
Cloud Computing is a new computing model consists of a large pool of hardware and software resources on remote datacenters that are accessed through the Internet. Cloud Computing faces significant obstacles to its acceptance, such as security, virtualization, and lack of standardization. For Cloud standards, there is a long debate about their role, and more demands for Cloud standards are put on the table. The Cloud standardization landscape is so ambiguous. To model and analyze security...
Show moreCloud Computing is a new computing model consists of a large pool of hardware and software resources on remote datacenters that are accessed through the Internet. Cloud Computing faces significant obstacles to its acceptance, such as security, virtualization, and lack of standardization. For Cloud standards, there is a long debate about their role, and more demands for Cloud standards are put on the table. The Cloud standardization landscape is so ambiguous. To model and analyze security standards for Cloud Computing and web services, we have surveyed Cloud standards focusing more on the standards for security, and we classified them by groups of interests. Cloud Computing leverages a number of technologies such as: Web 2.0, virtualization, and Service Oriented Architecture (SOA). SOA uses web services to facilitate the creation of SOA systems by adopting different technologies despite their differences in formats and protocols. Several committees such as W3C and OASIS are developing standards for web services; their standards are rather complex and verbose. We have expressed web services security standards as patterns to make it easy for designers and users to understand their key points. We have written two patterns for two web services standards; WS-Secure Conversation, and WS-Federation. This completed an earlier work we have done on web services standards. We showed relationships between web services security standards and used them to solve major Cloud security issues, such as, authorization and access control, trust, and identity management. Close to web services, we investigated Business Process Execution Language (BPEL), and we addressed security considerations in BPEL and how to enforce them. To see how Cloud vendors look at web services standards, we took Amazon Web Services (AWS) as a case-study. By reviewing AWS documentations, web services security standards are barely mentioned. We highlighted some areas where web services security standards could solve some AWS limitations, and improve AWS security process. Finally, we studied the security guidance of two major Cloud-developing organizations, CSA and NIST. Both missed the quality of attributes offered by web services security standards. We expanded their work and added benefits of adopting web services security standards in securing the Cloud.
Show less - Date Issued
- 2013
- PURL
- http://purl.flvc.org/fau/fd/FA0004001
- Subject Headings
- Cloud Computing, Computational grids (Computer systems), Computer network architectures, Expert systems (Computer science), Web services -- Management
- Format
- Document (PDF)
- Title
- Data mining heuristic-¬based malware detection for android applications.
- Creator
- Peiravian, Naser, Zhu, Xingquan, College of Engineering and Computer Science, Department of Computer and Electrical Engineering and Computer Science
- Abstract/Description
-
The Google Android mobile phone platform is one of the dominant smartphone operating systems on the market. The open source Android platform allows developers to take full advantage of the mobile operation system, but also raises significant issues related to malicious applications (Apps). The popularity of Android platform draws attention of many developers which also attracts the attention of cybercriminals to develop different kinds of malware to be inserted into the Google Android Market...
Show moreThe Google Android mobile phone platform is one of the dominant smartphone operating systems on the market. The open source Android platform allows developers to take full advantage of the mobile operation system, but also raises significant issues related to malicious applications (Apps). The popularity of Android platform draws attention of many developers which also attracts the attention of cybercriminals to develop different kinds of malware to be inserted into the Google Android Market or other third party markets as safe applications. In this thesis, we propose to combine permission, API (Application Program Interface) calls and function calls to build a Heuristic-Based framework for the detection of malicious Android Apps. In our design, the permission is extracted from each App’s profile information and the APIs are extracted from the packed App file by using packages and classes to represent API calls. By using permissions, API calls and function calls as features to characterize each of Apps, we can develop a classifier by data mining techniques to identify whether an App is potentially malicious or not. An inherent advantage of our method is that it does not need to involve any dynamic tracking of the system calls but only uses simple static analysis to find system functions from each App. In addition, Our Method can be generalized to all mobile applications due to the fact that APIs and function calls are always present for mobile Apps. Experiments on real-world Apps with more than 1200 malwares and 1200 benign samples validate the algorithm performance. Research paper published based on the work reported in this thesis: Naser Peiravian, Xingquan Zhu, Machine Learning for Android Malware Detection Using Permission and API Calls, in Proc. of the 25th IEEE International Conference on Tools with Artificial Intelligence (ICTAI) – Washington D.C, November 4-6, 2013.
Show less - Date Issued
- 2013
- PURL
- http://purl.flvc.org/fau/fd/FA0004045
- Subject Headings
- Computer networks -- Security measures, Data encryption (Computer science), Data structures (Computer science), Internet -- Security measures
- Format
- Document (PDF)
- Title
- Unifying the conceptual levels of network security through the use of patterns.
- Creator
- Kumar, Ajoy, Fernandez, Eduardo B., Florida Atlantic University, College of Engineering and Computer Science, Department of Computer and Electrical Engineering and Computer Science
- Abstract/Description
-
Network architectures are described by the International Standard for Organization (ISO), which contains seven layers. The internet uses four of these layers, of which three are of interest to us. These layers are Internet Protocol (IP) or Network Layer, Transport Layer and Application Layer. We need to protect against attacks that may come through any of these layers. In the world of network security, systems are plagued by various attacks, internal and external, and could result in Denial...
Show moreNetwork architectures are described by the International Standard for Organization (ISO), which contains seven layers. The internet uses four of these layers, of which three are of interest to us. These layers are Internet Protocol (IP) or Network Layer, Transport Layer and Application Layer. We need to protect against attacks that may come through any of these layers. In the world of network security, systems are plagued by various attacks, internal and external, and could result in Denial of Service (DoS) and/or other damaging effects. Such attacks and loss of service can be devastating for the users of the system. The implementation of security devices such as Firewalls and Intrusion Detection Systems (IDS), the protection of network traffic with Virtual Private Networks (VPNs), and the use of secure protocols for the layers are important to enhance the security at each of these layers.We have done a survey of the existing network security patterns and we have written the missing patterns. We have developed security patterns for abstract IDS, Behavior–based IDS and Rule-based IDS and as well as for Internet Protocol Security (IPSec) and Transport Layer Security (TLS) protocols. We have also identified the need for a VPN pattern and have developed security patterns for abstract VPN, an IPSec VPN and a TLS VPN. We also evaluated these patterns with respect to some aspects to simplify their application by system designers. We have tried to unify the security of the network layers using security patterns by tying in security patterns for network transmission, network protocols and network boundary devices.
Show less - Date Issued
- 2014
- PURL
- http://purl.flvc.org/fau/fd/FA00004132, http://purl.flvc.org/fau/fd/FA00004132
- Subject Headings
- Computer architecture, Computer network architectures, Computer network protocols, Computer network protocols, Computer networks -- Security measures, Expert systems (Computer science)
- Format
- Document (PDF)