NEW CABLE CONVERTER FOR SEISMOMAGNETOMETRIC DETECTION.
ALEXEY ALEXEYEVICH TERENIN, PH.D. (Eng.)
CRYPTOGRAPHIC ALGORITHMS USED TO ENSURING INFORMATION SECURITY DURING INTERNET INTERACTION A brief overview of the most common encryption algorithms today, their description, as well as problems that arise during their implementation and significant aspects in practical use are presented. Protection of information by cryptographic transformation methods consists of changing its constituent parts (words, letters, syllables, numbers) using special algorithms or hardware solutions and key codes, i.e., bringing it to an implicit form. To get acquainted with encrypted information, the reverse process is used: decoding (decryption). The use of cryptography is one of the common methods that significantly increases the security of data transmission in computer networks, data stored in remote memory devices, as well as when exchanging information between remote objects. For transformation (encryption), some algorithm or device implementing a given algorithm, which may be known to a wide range of people, is usually used. The encryption process is controlled by a periodically changing key code, which ensures an original presentation of information each time when using the same algorithm or device. Knowing the key allows you to decrypt the text simply and reliably. However, without knowing the key, this procedure may be practically impossible even with a known encryption algorithm. Even simple transformation of information is a very effective means of hiding its meaning from most unskilled intruders.
A Brief Historical Review of the Development of Encryption The origins of cryptography go back to Egyptian hieroglyphs. Since ancient times, when Egypt and Persia flourished, messengers were used for the most important state and military missions, who carried the text of the message either on parchment or in their heads to convey it verbally, the latter method was more preferable. Even then, more or less successful methods of protecting transmitted information from the encroachments of interceptors appeared. Let us cite a well-known legend from the Ancient World. A certain king, having been captured, made a tattoo on the head of a slave — a message to the allies. When the hair grew back, the slave moved to the recipients of the message and the king was freed. The prototype of today's steganography. The ancient Greeks used round sticks of the same diameter, on which strips of parchment were wound. The writing was done longitudinally along the length of the stick. It was possible to fold the text into a readable form only if you had a stick of the same diameter. In Ancient Rome, the science of cryptography, translated from Latin as secret writing, was already clearly beginning to form. The Caesar cipher appears, when each letter is replaced by a letter three letters away in the alphabet. In medieval intriguing Europe and Central Asia, there was a rapid development of cryptography and cryptanalysis — methods for breaking encrypted texts. The first systematic work on cryptography is considered to be the book of the architect Leon Battisti Alberti (1404 — 1472). One of the first cryptanalysts was François Viète (1540 — 1603), at the court of the King of France Henry IV. At the same time, advisers from the Adgenti family, who can also be called cryptanalysts, served at the court of the Pope. The entire period up to the middle of the 17th century is full of works on cryptography and cryptanalysis. In the 19th and in the first half of the 20th century, many countries, including Russia, used encryption methods for secret diplomatic correspondence, the keys for which were composed of excerpts from certain texts of ordinary books (cipher books). Since the beginning of the 20th century, — since the First World War — special encryption machines begin to be used. The German Enigma machine, the code of which was broken by the British, is widely known. In order not to reveal the fact that the German cipher had been broken, the British government made great sacrifices among the civilian population by not warning the residents of two large cities about the impending bombings. But this later helped to gain a significant advantage in the northern naval battles with Germany, when the invincible German submarines and cruisers were destroyed. After World War II, computers took up cryptography. For a long time, it was the domain of the most powerful supercomputers of their time. Publications on this topic were strictly classified and the use of scientific research in this area was a national prerogative. Only the textbook work of Von Neumann in the 1940s was publicly available, describing, in addition to the principles of building computing systems, some other possible malicious methods of influence to disrupt the “legal” computing process, as well as the classic work of Shannon, which laid the foundations of computer cryptography. Since the 1970s, open publications have appeared: Heffey-Dillman in 1976. In 1970, there was a classified invention of James Ellis (Great Britain) in the field of cryptography. The most famous algorithm of asymmetric cryptography is RSA, developed by Ronald Rivest, Eddie Shamir and Len Edleman in 1977. The RSA algorithm is of great importance because can be used for both public key encryption and digital signature creation. This was a revolutionary period in the development of cryptographic science. New methods of secret distribution of key information in open computing systems appeared, and asymmetric cryptography was born. But even after that, for a long time, the prerogative of using cryptography in data protection was with government services and large corporations. The computing equipment of that time, which had the power necessary for cryptographic transformations, was very expensive. At that time, the main state standards of cryptographic algorithms appeared (USA and some European countries), the use of which was prescribed when working with information classified as a state secret. The veil of secrecy surrounding these technologies even led to the US equating cryptographic algorithms with weapons, and a ban on the export of encryption hardware and software. Then, export restrictions were introduced on the length of the key used in encryption algorithms outside the US, which allowed American intelligence agencies to decrypt messages using existing computing power without knowing the shortened key. Export restrictions were lifted on March 1, 2001. Due to the events of September 11 of the same year, there has been a tightening of government control. The US government is considering options for re-introducing export controls on encryption tools. Let's go back to the 70s. Since then, neither scientific research nor the development of computing tools have stopped. The computing power of supercomputers increases several times every few years. The personal computer appears. The power of a personal computer is approximately equal to the power of a supercomputer ten years ago. Now personal computers have become even more powerful. Since the 1980s, ordinary users have had the opportunity to use cryptographic tools on their computers, which government agencies have been fiercely preventing, and it has become more difficult to monitor the activities of citizens of the country, including criminals. The release of PGP (Pretty Good Privacy) by Phil Zimmermann (version 1.0 was released in 1991) and its provision for open and free use gave great opportunities to ordinary computer users. Phil Zimmermann was even declared an enemy of the state, he was sentenced to imprisonment. Ever-increasing computing power forced the use of increasingly complex cryptographic algorithms or increasing the length of the keys used in encryption. Standards for cryptographic algorithms became obsolete and unreliable. Information protected by a certain key could no longer be kept confidential for a long enough time — as long as required by government regulations. For example, keeping information completely secret in encrypted form for 5 years meant that an adversary with the most powerful computing resources, constantly trying out possible keys, would not have been able to find the right key to decrypt the stored information within that period with a fairly high probability. Competitions began to be held to crack some information encrypted using the algorithm of one of the standards. The winner was awarded a solid cash prize, as well as worldwide fame in the information community. By combining ordinary computers in a computer network to work in parallel to solve a given problem, users gathered in groups and selected the key together. A key length of 48 bits means that 248 attempts must be made. Increasing the key length, for example, by only 16 bits, means that 216 times more must be tried. But even such a key size allowed the problem of cracking the cipher to be solved by the combined groups in days and even hours of parallel work. Later, it was necessary to switch to keys that were several times longer than those mentioned. But this was only a temporary measure, and new standards for cryptographic transformation algorithms (AES in the USA) were recently adopted. Currently, the press has published many publications devoted to this problem. Numerous books are published, both translated and by Russian authors. Cryptography allows solving the problem of protecting information from disclosure and modification. The mathematical apparatus of modern cryptography surpasses in complexity that used to develop nuclear weapons and space systems. Modern cryptography is divided into symmetric and asymmetric. Symmetric — into stream cipher, block and composite. Asymmetric cryptography is more resource-intensive, and in symmetric there is a problem of effective key distribution. Modern secure exchange systems are based on the use of mixed cryptography. At the beginning of the exchange session, the parties send each other secret session keys using asymmetric cryptography, which are then used for symmetric encryption of the transmitted data. The asymmetric cryptography system allows distributing keys in symmetric encryption systems. Government and military telecommunications systems use exclusively symmetric encryption (usually with one-time keys). This is due to the fact that the stability of systems with public keys has not been proven strictly mathematically, but the opposite has not been proven either. Encryption of information should not be taken as a panacea for all information threats. It should be perceived as one of the mandatory measures to protect information as part of a comprehensive information security system. The use of encryption should be combined with legislative, organizational and other protective measures.
Symmetric encryption algorithms Encryption algorithms are designed to solve the problem of ensuring the confidentiality of information. Currently, cryptographic methods are intensively used to close information. Since ancient times, the most effective form of protection has been and remains encryption. Encryption is defined as the mutual conversion of unprotected (open) information into encrypted (closed) form – ciphertext [1], in which it is not presented fully accessible to an attacker. Keys are used during encryption, the presence of which means the possibility of encryption and/or decryption of information. It is important to note that the encryption method itself does not need to be kept secret, since knowledge of it alone will not allow decryption of the ciphertext. Modern cryptosystems can be clearly divided by the method of using keys into cryptosystems with a secret key (symmetric) and with an open key (asymmetric). If the same key is used for encryption and decryption, such a cryptosystem is called symmetric. Symmetric cryptosystems include DES [2], AES, GOST 28147-89 [3], etc. A new direction in cryptography was the invention of asymmetric cryptosystems with an open key, such as RSA, DSA or El-Gamal [4, 5]. In asymmetric cryptosystems, different keys are used for encryption and decryption, which are practically inderivable from each other, one of which (the decryption key) is made secret, and the other (the encryption key) is made public. This makes it possible to transmit secret messages over an unprotected channel without first transmitting the secret key. It was public key cryptography that broke the vicious circle of symmetric ciphers, when in order to organize the exchange of secret information, it was first necessary to distribute secret keys. Public key cryptosystems will be considered in detail later, but now let us return to symmetric cryptosystems (KS). The most important component of KS are ciphers [1, 4] or procedures for the mutually inverse transformation of plaintext M into ciphertext M': M’ = E(M), where E is the encryption function and D is the decryption function. The generally accepted approach in cryptography is to construct a cipher in which its secrecy is determined only by the secrecy of the KS key (Kerkoff's rule). Thus, the cipher must be resistant to cracking, even if a potential cryptanalyst knows the entire encryption algorithm, except for the value of the key used, and has the full text of the intercepted ciphertext. Practice has shown that the more the algorithm is known, the more people have worked with it, the more tested, and therefore reliable it becomes. Thus, publicly known algorithms now withstand the struggle with time, but classified state ciphers reveal many errors and shortcomings, since it is impossible to take everything into account. The generally accepted scheme for constructing symmetric cryptosystems is cyclic permutations and substitutions of bits in a block of fixed length, the algorithm of which is determined by the secret key.
An encryption algorithm is considered secure if, having private data and knowing the secret key, it is impossible to obtain information about the open data. It has been rigorously proven that it is impossible to construct an absolutely secure cipher, except for the case when the size of the secret key is equal to (or greater than) the size of the encrypted data [1]. This case is difficult to implement in practice, since the cryptographic protection tools actually used and available on the market use ciphers for which the task of restoring the open text from the closed text is difficult to calculate, that is, it requires such large resources that the attack becomes economically impractical. Among the symmetric ciphers, the following are the most well-known and frequently used (the block size in bits is designated as b, the number of cycles is r, and the key length is l): DES— US government standard [2] (b = 64, r = 16, l = 56). Currently, DES has been proven to be insufficiently resistant to brute force attacks [6, 7]. Now the new encryption standard AES has been adopted in the USA. A competition was held among encryption algorithms, in which Rijndael won and became the basis for AES. Rijndael is an iterative block cipher with variable block length and different key lengths. A more detailed description of this algorithm and the results of the competition are given in [18]. A fairly large number of symmetric algorithms have been developed, published and studied in the world (Table 1), of which only DES and its modification Triple DES have been sufficiently tested by time. The table does not include little-known and poorly studied algorithms, such as Safer, etc.
Table 1. Overview of symmetric encryption methods
Currently, symmetric algorithms with a key length of more than 100 bits (Triple DES and IDEA, etc.) are not unbreakable. The domestic GOST algorithm, in comparison with them, is distinguished by increased complexity both in generating replacement nodes and in generating keys. Also, for the GOST algorithm, there is a high probability of generating an unstable key, which in some encryption modes reduces its effective key length from 2256 to 262 [12]. Triple DES is a more proven algorithm than IDEA and provides acceptable operating speed. The Triple DES algorithm is a three-fold application of the DES algorithm to the same data, but with different keys. DES has penetrated into Russia and is widely used in practice as an integral part of various software and hardware, the most widely known of which are the S.W.I.F.T. system, secret VISA and EUROPAY modules, secret ATM and POS modules, and, finally, smart cards. Particularly intense discussions around data encryption algorithms are caused by smart cards. At the same time, there are serious grounds to believe that the reliability of domestic cryptosystems of conversion origin will exceed foreign analogues [16]. However, Russian legislation, as well as the legislation of many other countries, only permits the use of national encryption standards. The GOST 28147-89 algorithm is built on the same principle as DES, it is a classic block cipher with a secret key, but differs from DES in a longer key length, a larger number of rounds and a simpler scheme for constructing the rounds themselves. Table 2 shows its main parameters, for convenience — in comparison with the parameters of DES [17].
Table 2. Comparison of the parameters of the DES and GOST ciphers
If secret information needs to be exchanged between people who trust each other, i.e. who are part of the same organization, symmetric cryptography can be used. Of course, both (or more) parties must already have encryption keys for interaction. If we briefly describe the information exchange scenario, it is as follows:
Many modern methods of protective transformations can be classified into four large groups: permutations, replacements (substitutions), additive and combined methods. Permutation and substitution methods are usually characterized by a short key length, and the reliability of their protection is determined by the complexity of the transformation algorithms. Additive methods are characterized by simple transformation algorithms, and their cryptographic resistance is based on increasing the key length.
Cipher cracking There is a method for cracking a cipher based on trying all the key options. The criterion for the correctness of the option is the presence of the probable word in the text”. The set of all possible keys is searched, the encrypted text is decrypted on each key. A probable word is searched for in the resulting “pseudo-open” text. If there is no such word, the current text is rejected, and the next key is selected. If such a word is found, the key variant is displayed on the screen. Then the keys are searched until the entire set of variants is exhausted. It is possible to find several keys for which the “pseudo-open” texts contain a probable word. After the search is complete, it is necessary to decrypt the text on the keys found. The pseudo-open text is displayed on the screen for visual control. If the operator recognizes the text as open, then the work on opening is finished. Otherwise, this key variant is rejected and the transition to the next key is carried out. To combat the method of complete search, it is possible to increase the length of the encryption key used. Moreover, increasing its length by only 8 bits increases the number of search variants by 28 times, respectively by 64 bits — by 264 times. Among the problems inherent in the use of cryptographic encryption algorithms, it is necessary to highlight the problem of key distribution. Before the interacting parties can send each other encrypted messages, they must exchange encryption keys via some secret channel. In addition, the information exchange system must maintain a huge number of keys in an up-to-date state. Cryptographic encryption algorithms do not allow to establish the integrity of the received message (i.e. to make sure that the message was not modified during transmission). Authorship can be confirmed only by possession of a certain key, so anyone who becomes the owner of someone else's key will be able to pass off their messages as messages sent by another user. The problem of distributing secret keys over a public communication channel can be solved by the Diffie-Hellman algorithm. But this algorithm is an asymmetric cryptographic algorithm. They use two keys: public and private. Asymmetric cryptographic algorithms were rapidly developed in the 1970s. Such algorithms can also solve the problems of authorship and authenticity verification, allowing encrypted information to be exchanged between parties that do not trust each other. In addition, the use of asymmetric algorithms reduces by an order of magnitude the number of keys that must be distributed between the interacting parties. Asymmetric encryption systems include a public database of public keys that can be distributed over open communication channels and their disclosure will in no way compromise the system, which is why they are called open. End in No. 4, 2006.
Literature
|