All papers in 2008 (Page 3 of 545 results)
An Efficient Authenticated Key Exchange Protocol with a Tight Security Reduction
In this paper, we present a new authenticated key exchange(AKE)
protocol, called NETS, and prove its security in the extended
Canetti-Krawczyk model under the random oracle assumption and the
gap Diffie-Hellman(GDH) assumption. Our protocol enjoys a simple and
tight security reduction compared to those of HMQV and CMQV without
using the Forking Lemma. Each session of the NETS protocol requires
only three exponentiations per party, which is comparable to the
efficiency of MQV, HMQV and CMQV.
Authenticated Key Exchange Secure under the Computational Diffie-Hellman Assumption
In this paper, we present a new authenticated key exchange(AKE)
protocol and prove its security under the random oracle assumption
and the computational Diffie-Hellman(CDH) assumption. In the
extended Canetti-Krawczyk model, there has been no known AKE
protocol based on the CDH assumption. Our protocol, called NAXOS+,
is obtained by slightly modifying the NAXOS protocol proposed by
LaMacchia, Lauter and Mityagin. We establish a formal security proof
of NAXOS+ in the extended Canetti-Krawczyk model using as a main
tool the trapdoor test presented by Cash, Kiltz and Shoup.
Efficient RFID authentication protocols based on pseudorandom sequence generators
Uncategorized
Uncategorized
In this paper, we introduce a new class of PRSGs, called partitioned pseudorandom sequence generators(PPRSGs), and propose an RFID authentication protocol using a PPRSG, called S-protocol. Since most existing stream ciphers can be regarded as secure PPRSGs, and stream ciphers outperform other types of symmetric key primitives such as block ciphers and hash functions in terms of power, performance and gate size, S-protocol is expected to be suitable for use in highly constrained environments such as RFID systems. We present a formal proof that guarantees resistance of S-protocol to desynchronization and tag-impersonation attacks. Speci¯cally, we reduce the availability of S-protocol to pseudorandomness of the underlying PPRSG, and the security of the protocol to the availability. Finally, we give a modi¯cation of S-protocol, called S¤-protocol, that provides mutual authentication of tag and reader.
Cryptanalysis of Li et al.'s Identity-Based Threshold Signcryption Scheme
Signcryption is a cryptographic primitive that aims at providing confidentiality and authentication simultaneously. Recently in May
2008, a scheme for identity based threshold signcryption was
proposed by Fagen Li and Yong Yu. They have proved the
confidentiality of their scheme and have also claimed the
unforgeability without providing satisfactory proof. In this paper,
we show that in their signcryption scheme the secret key of the
sender is exposed(total break) to the clerk during sincryption and
hence insecure in the presence of malicious clerks. Further, we
propose a corrected version of the scheme and formally prove its
security under the existing security model for signcryption.
An Efficient Identity-Based Signcryption Scheme for Multiple Receivers
This paper puts forward a new efficient construction for Multi-Receiver Signcryption
in the Identity-based setting. We consider a scenario where a user wants to securely send a message
to a dynamically changing subset of the receivers in such a way that non-members of the of this
subset cannot learn the message. The obvious solution is to transmit an individually signcrypted
message to every member of the subset. This requires a very long transmission (the number of
receivers times the length of the message) and high computation cost. Another simple solution
is to provide every possible subset of receivers with a key. This requires every user to store a
huge number of keys. In this case, the storage efficiency is compromised. The goal of this paper
is to provide solutions which are efficient in all three measures i.e. transmission length, storage of
keys and computation at both ends. We propose a new scheme that achieve both confidentiality
and authenticity simultaneously in this setting and is the most efficient scheme to date, in the
parameters described above. It breaks the barrier of ciphertext length of linear order in the number
of receivers, and achieves constant sized ciphertext, independent of the size of the receiver set. This
is the first Multi-receiver Signcryption scheme to do so. We support the scheme with security proofs
under a precisely defined formal security model
Last updated: 2011-12-05
On construction of signature schemes based on birational permutations over noncommutative rings
In the present paper, we give a noncommutative version of Shamir's birational permutation signature scheme proposed in Crypto'93 in terms of square matrices. The original idea to construct the multivariate quadratic signature is to hide a quadratic triangular system using two secret linear transformations. However, the weakness of the triangular system remains even after taking two transformations, and actually Coppersmith et al. broke it linear algebraically. In the non-commutative case, such linear algebraic weakness does not appear. We also give several examples of noncommutative rings to use in our scheme, the ring consisting of all square matrices, the quaternion ring and a subring of three-by-three matrix ring generated by the symmetric group of degree three.
Note that the advantage of Shamir's original scheme is its efficiency. In our scheme, the efficiency is preserved enough.
High Performance Implementation of a Public Key Block Cipher - MQQ, for FPGA Platforms
We have implemented in FPGA recently published class of public key algorithms -- MQQ, that are based on quasigroup string transformations. Our implementation achieves decryption throughput of 399 Mbps on an Xilinx Virtex-5 FPGA that is running on 249.4 MHz. The encryption throughput of our implementation achieves 44.27 Gbps on four Xilinx Virtex-5 chips that are running on 276.7 MHz. Compared to RSA implementation on the same FPGA platform this implementation of MQQ is 10,000 times faster in decryption, and is more than 17,000 times faster in encryption.
An improvement of discrete Tardos fingerprinting codes
Uncategorized
Uncategorized
It has been known that the code lengths of Tardos's collusion-secure fingerprinting codes are of theoretically minimal order with respect to the number of adversarial users (pirates). However, the code lengths can be further reduced, as some preceding studies on Tardos's codes already revealed. In this article we improve a recent discrete variant of Tardos's codes, and give a security proof of our codes under an assumption weaker than the original assumption (Marking Assumption).
Our analysis shows that our codes have significantly shorter lengths than Tardos's codes. For example, in a practical setting, the code lengths of our codes are about 3.01%, 4.28%, and 4.81% of Tardos's codes if the numbers of pirates are 2, 4, and 6, respectively.
Modified Huang-Wang's Convertible Nominative Signature Scheme
At ACISP 2004, Huang and Wang first introduced the concept of
convertible nominative signatures and also proposed a concrete
scheme. However, it was pointed out by many works that Huang-Wang's
scheme is in fact not a nominative signature. In this paper, we
first present a security model for convertible nominative
signatures. The properties of Unforgeability, Invisibility,
Non-impersonation and Non-repudiation in the setting of convertible
nominative signatures are defined formally. Then we modify
Huang-Wang's scheme into a secure one. Formal proofs are provided to
show that the modified Huang-Wang's scheme satisfies all the
security properties under some conventional assumptions in the
random oracle model.
New attacks on ISO key establishment protocols
Cheng and Comley demonstrated type flaw attacks against the key establishment mechanism 12 standardized in ISO/IEC 11770-2:1996. They also proposed enhancements to fix the security flaws in the mechanism. We show that the enhanced version proposed by Cheng and Comley is still vulnerable to type flaw attacks. As well we show that the key establishment mechanism 13 in the above standard is vulnerable to a type flaw attack.
Public Key Cryptography from Different Assumptions
We construct a new public key encryption based on two
assumptions:
1) One can obtain a pseudorandom generator with small locality by connecting the outputs to the inputs using any sufficiently good unbalanced expander.
2) It is hard to distinguish between a random graph that is such an expander and a random graph where a (planted) random logarithmic-sized subset S of the outputs is connected to fewer than |S| inputs.
The validity and strength of the assumptions raise interesting new algorithmic and pseudorandomness questions, and we explore their relation to the current state-of-art.
Analyzing the Galbraith-Lin-Scott Point Multiplication Method for Elliptic Curves over Binary Fields
Galbraith, Lin and Scott recently constructed efficiently-computable
endomorphisms for a large family of elliptic curves defined over
F_{q^2} and showed, in the case where q is prime, that the
Gallant-Lambert-Vanstone point multiplication method for these curves
is significantly faster than point multiplication for general elliptic
curves over prime fields. In this paper, we investigate the potential
benefits of using Galbraith-Lin-Scott elliptic curves in the case
where q is a power of 2. The analysis differs from the q prime case
because of several factors, including the availability of the point
halving strategy for elliptic curves over binary fields. Our analysis
and implementations show that Galbraith-Lin-Scott offers significant
acceleration for curves over binary fields, in both doubling- and
halving-based approaches. Experimentally, the acceleration surpasses
that reported for prime fields (for the platform in common), a
somewhat counterintuitive result given the relative costs of point
addition and doubling in each case.
Explicit hard instances of the shortest vector problem
Building upon a famous result due to Ajtai, we propose a sequence of lattice bases with growing dimension, which can be expected to be hard instances of the shortest vector problem (SVP) and which can therefore be used to benchmark lattice reduction algorithms.
The SVP is the basis of security for potentially post-quantum cryptosystems.
We use our sequence of lattice bases to create a challenge, which may be helpful in determining appropriate parameters for these schemes.
Efficient Key Distribution Schemes for Large Scale Mobile Computing Applications
In emerging networks consisting of large-scale deployments of mobile devices, efficient security mechanisms are required to facilitate cryptographic authentication. While computation and bandwidth overheads are expensive for mobile devices, the cost of storage resources continue to fall at a rapid rate. We propose a simple novel key predistribution scheme, \textit{key subset and symmetric certificates} (KSSC) which can take good advantage of inexpensive storage resources, and has many compelling advantages over other approaches for facilitating ad hoc establishment of pairwise secrets in mobile computing environments. We argue that a combination of KSSC with a variant of an elegant KDS proposed by Leighton and Micali is an appealing choice for securing large scale deployments of mobile devices.
A Secure Remote User Authentication Scheme with Smart Cards
Remote user authentication scheme is one of the
simplest and the most convenient authentication mechanisms
to deal with secret data over insecure networks.
These types of schemes are applicable to the areas
such as computer networks, wireless networks, remote
login systems, operation systems and database management
systems.The goal of a remote user authentication
scheme is to identify a valid card holder as having
the rights and privileges indicated by the issuer of
the card. In recent years, so many remote user authentication
schemes have been proposed to authenticate a
legitimate user, but none of them can solve all possible
problems and withstand all possible attacks. This
paper presents a secure remote user authentication
scheme with smart cards. The proposed scheme provides
the essential security requirements and achieves
particular attributes.
Last updated: 2009-10-23
Chosen ciphertext secure public key encryption under DDH assumption with short ciphertext
Uncategorized
Uncategorized
An efficient variant of the ElGamal public key encryption scheme is proposed which is
provably secure against adaptive chosen ciphertext attacks(IND-CCA2) under the decisional
Diffie-Hellman(DDH) assumption. Compared to the previously most efficient scheme under
DDH assumption by Kurosawa and Desmedt [Crypto 2004] it has one group element shorter
ciphertexts, 50\% shorter secret keys, 25\% shorter public keys and it is 28.6\% more
efficient in terms of encryption speed, 33.3\% more efficient in terms of decryption
speed. A new security proof logic is used, which shows directly that the decryption
oracle will not help the adversary in the IND-CCA2 game. Compared to the previous
security proof, the decryption simulation is not needed in the new logic. This makes the
security proof simple and easy to understand.
SMS4 Encryption Algorithm for Wireless Networks
SMS4 is a Chinese block cipher standard, mandated for use in protecting wireless net-
works, and issued in January 2006. The input, output, and key of SMS4 are each 128 bits.
The algorithm has 32 rounds, each of which modifies one of the four 32-bit words that make
up the block by xoring it with a keyed function of the other three words. Encryption and
decryption have the same structure except that the round key schedule for decryption is the
reverse of the round key schedule for encryption.
Attribute-Based Signatures: Achieving Attribute-Privacy and Collusion-Resistance
Uncategorized
Uncategorized
We introduce a new and versatile cryptographic primitive called {\em Attribute-Based Signatures} (ABS), in which a signature attests not to the identity of the individual who endorsed a message, but instead to a (possibly complex) claim regarding the attributes she posseses. ABS offers:
* A strong unforgeability guarantee for the verifier,
that the signature was produced by a {\em single} party whose
attributes satisfy the claim being made; i.e., not by a
collusion of individuals who pooled their attributes together.
* A strong privacy guarantee for the signer, that the
signature reveals nothing about the identity or attributes of the
signer beyond what is explicitly revealed by the claim being made.
We formally define the security requirements of ABS as a cryptographic primitive, and then describe an efficient ABS construction based on groups with bilinear pairings. We prove that our construction is secure in the generic group model. Finally, we illustrate several applications of this new tool; in particular, ABS fills a critical security requirement in attribute-based messaging (ABM) systems.
A powerful feature of our ABS construction is that unlike many other attribute-based cryptographic primitives, it can be readily used
in a {\em multi-authority} setting, wherein users can make claims involving combinations of attributes issued by independent
and mutually distrusting authorities.
Blind HIBE and its Applications to Identity-Based Blind Signature and Blind Decryption
We explicitly describe and analyse \textit{blind} hierachical identity-based encryption (\textit{blind} HIBE) schemes, which are natural generalizations of blind IBE schemes \cite{gh07}.
We then uses the blind HIBE schemes to construct: (1) An identity-based blind signature scheme secure in the standard model, under the computational Diffie-Hellman (CDH) assumption, and with much shorter signature size and lesser communication cost, compared to existing proposals. (2) A new mechanism supporting a user to buy digital information over the Internet without revealing what he/she has bought, while protecting the providers from cheating users.
Two attacks on a sensor network key distribution scheme of Cheng and Agrawal
A sensor network key distribution scheme
for hierarchical sensor networks was recently proposed by
Cheng and Agrawal. A feature of their scheme is that
pairwise keys exist between any pair of high-level nodes
(which are called cluster heads) and between any (low-level)
sensor node and the nearest cluster head. We present two attacks on their scheme.
The first attack can be applied for certain
parameter sets. If it is applicable, then this attack can result in the
compromise of most if not all of
the sensor node keys after a small number of cluster heads are compromised.
The second attack can always be applied, though it is weaker.
Revisit of Group-based Unidirectional Proxy Re-encryption Scheme
Uncategorized
Uncategorized
Currently, researchers have focused their attention on proxy re-encryption scheme deployed between two entities. Lots of bidirectional schemes have been proposed and this kind of scheme is suitable for the scenario in which the two entities have already established a relationship of trust. How to construct a unidirectional scheme is an open problem and receiving increasing attention. In this paper, we present a unidirectional proxy re-encryption scheme for group communication. In this scheme, a proxy is only allowed to convert ciphertext for Alice into ciphertext for Bob without revealing any information on plaintext or private key. It is suitable for the environment in which no mutual relationship exists and transitivity is not permitted. We prove the scheme secure against chosen ciphertext attack in standard model.
RSA-TBOS Signcryption with Proxy Re-encryption.
Uncategorized
Uncategorized
The recent attack on Apple iTunes Digital Rights Management \cite{SJ05} has brought to light the usefulness of proxy re-encryption schemes for Digital Rights Management. It is known that the use of proxy re-encryption would have prevented the attack in \cite{SJ05}. With this utility in mind and with the added requirement of non-repudiation, we propose the first ever signcryption scheme with proxy re-encryption that does not involve bilinear maps. Our scheme is called RSA-TBOS-PRE and is based on the RSA-TBOS signcryption scheme of Mao and Malone-Lee \cite{MM03}. We adapt various models available in the literature concerning authenticity, unforgeability and non-repudiation and propose a signature non-repudiation model suitable for signcryption schemes with proxy re-encryption. We show the non-repudiability of our scheme in this model. We also introduce and define a new security notion of Weak-IND-CCA2, a slightly weakened adaptation of the IND-CCA2 security model for signcryption schemes and prove that RSA-TBOS-PRE is secure in this model. Our scheme is Weak-IND-CCA2 secure, unidirectional, extensible to multi-use and does not use bilinear maps. This represents significant progress towards solving the open problem of designing an IND-CCA2 secure, unidirectional, multi-use scheme not using bilinear maps proposed in \cite{CH07}\cite{SXC08}.
A new identity based proxy signature scheme
Proxy signature schemes allow a proxy signer to generate proxy signatures on behalf of an original signer. Mambo et al. first introduced the notion of proxy signature and a lot of research work can be found on this topic nowadays. Recently, many identity based proxy signature schemes were proposed. However, some schemes are vulnerable to proxy key exposure attack. In this paper, we propose a security model for identity based proxy signature schemes. Then an efficient scheme from pairings is presented. The presented scheme is provably secure in the random oracle model. In particular, the new scheme is secure against proxy key exposure attack.
Lattice-based Blind Signatures
Uncategorized
Uncategorized
Blind signatures (BS), introduced by Chaum, have become a cornerstone in privacy-oriented cryptography.
Using hard lattice problems, such as the shortest vector problem, as the basis of security has advantages over using the factoring or discrete logarithm problems. For instance, lattice operations are more efficient than modular exponentiation and lattice problems remain hard for quantum and subexponential-time adversaries.
Generally speaking, BS allow a signer to sign a message without seeing it, while retaining a certain amount of control over the process. In particular, the signer can control the number of issued signatures. For the receiver of the signature, this process provides perfect anonymity, e.g., his spendings remain anonymous when using BS for electronic money.
We provide a positive answer to the question of whether it is possible to implement BS based on lattice problems. More precisely, we show how to turn Lyubashevsky's identification scheme into a BS scheme, which has almost the same efficiency and security in the random oracle model. In particular, it offers quasi-linear complexity, statistical blindness, and its unforgeability is based on the hardness of worst-case lattice problems with an approximation factor of in dimension . Moreover, it is the first blind signature scheme that supports leakage-resilience, tolerating leakage of a fraction of the secret key in a model that is inspired by Katz and Vaikuntanathan.
A correction to ``Efficient and Secure Comparison for On-Line Auctions''
In this note, we describe a correction to the cryptosystem proposed by the authors in "Efficient and Secure Comparison for On-Line Auctions". Although the correction is small and does not affect the performance of
the protocols proposed in that paper, it is necessary as the cryptosystem is not secure without it.
Public Key Block Cipher Based on Multivariate Quadratic Quasigroups
Uncategorized
Uncategorized
We have designed a new class of public
key algorithms based on quasigroup string transformations using a
specific class of quasigroups called \emph{multivariate quadratic
quasigroups (MQQ)}. Our public key algorithm is a bijective mapping,
it does not perform message expansions and can be used both for
encryption and signatures. The public key consist of quadratic
polynomials with variables where . A
particular characteristic of our public key algorithm is that it is
very fast and highly parallelizable. More concretely, it has the
speed of a typical modern symmetric block cipher -- the reason for
the phrase \emph{"A Public Key Block Cipher"} in the title of this
paper. Namely the reference C code for the 160--bit variant of the
algorithm performs decryption in less than 11,000 cycles (on Intel
Core 2 Duo -- using only one processor core), and around 6,000
cycles using two CPU cores and OpenMP 2.0 library. However,
implemented in Xilinx Virtex-5 FPGA that is running on 249.4 MHz it
achieves decryption throughput of 399 Mbps, and implemented on four
Xilinx Virtex-5 chips that are running on 276.7 MHz it achieves
encryption throughput of 44.27 Gbps. Compared to fastest RSA
implementations on similar FPGA platforms, MQQ algorithm is more
than 10,000 times faster.
Yet Another Secure Distance-Bounding Protocol
Distance-bounding protocols have been proposed by Brands and Chaum in 1993
in order to detect \emph{relay attacks}, also known as \emph{mafia fraud}.
Although the idea has been introduced fifteen years ago, only recently distance-bounding protocols
attracted the attention of the researchers.
Several new protocols have been proposed the last five years.
In this paper, a new secure distance-bounding protocol is presented. It is self-contained and composable
with other protocols for example for authentication or key-negotiation. It allows periodically execution
and achieves better use of the communication channels by exchanging authenticated nonces.
The proposed protocol becomes suitable for wider class of devices, since the resource
requirements to the prover are relaxed.
Attacking and defending the McEliece cryptosystem
This paper presents several improvements to Stern's attack on the McEliece cryptosystem and achieves results considerably better than Canteaut et al. This paper shows that the system with the originally proposed parameters can be broken in just 1400 days by a single 2.4GHz Core 2 Quad CPU,or 7 days by a cluster of 200 CPUs. This attack has been implemented and is now in progress.
This paper proposes new parameters for the McEliece and Niederreiter cryptosystems achieving standard levels of security against all known attacks. The new parameters take account of the improved attack; the recent introduction of list decoding for binary Goppa codes; and the possibility of choosing code lengths that are not a power of 2. The resulting public-key sizes are considerably smaller than previous parameter choices for the same level of security.
Last updated: 2010-02-09
Elliptic Curves Scalar Multiplication Combining Multi-base Number Representation with Point halving
Elliptic curves scalar multiplication over some finite fields, attractive research area, which paid much attention by researchers in the recent years. Researchs still in progress to improve elliptic curves cryptography implementation and reducing its complexity. Elliptic curve point-halving algorithm proposed in and later double-base chain and step multi-base chain are among efficient techniques offered in this field. Our paper proposes new algorithm combining step multi-base number representation and point halving. We extend the work done by K. W. Wong, which combined double base chain with point halving technique. The expriment results show our contribution will enhance elliptic curves scalar multiplication.
Signing a Linear Subspace: Signature Schemes for Network Coding
Network coding offers increased throughput and improved robustness
to random faults in completely decentralized networks.
In contrast to traditional routing schemes, however, network coding
requires intermediate nodes to modify data packets en route;
for this reason, standard signature schemes are inapplicable and it
is a challenge to provide resilience to tampering by malicious
nodes.
Here, we propose two signature schemes that can be used in
conjunction with network coding to prevent malicious modification of
data. In particular, our schemes can be viewed as signing linear
subspaces in the sense that a signature on V
authenticates exactly those vectors in V.
Our first scheme is homomorphic and has better performance,
with both public key size and per-packet overhead being constant.
Our second scheme does not rely on random oracles and uses weaker assumptions.
We also prove a lower bound on the length of signatures for
linear subspaces showing that both of our schemes are essentially optimal in
this regard.
RSA Cryptanalysis with Increased Bounds on the Secret Exponent using Less Lattice Dimension
We consider RSA with , , public encryption exponent and private decryption exponent . Boneh and Durfee (Eurocrypt 1999, IEEE-IT 2000) used Coppersmith's method (Journal of Cryptology, 1997) to factorize using when , the {\sf theoretical bound}. Related works have also been presented by Blömer and May (CaLC 2001). However, the {\sf experimental bound} for that has been reached so far is only for 1000 bits (the upper bound on less for higher number of bits). The basic idea relied on LLL algorithm, but the experimental bounds were constrained by large lattice dimensions. In this paper we present
{\sf theoretical results} as well as {\sf experimental evidences} to
{\sf extend the bound of} for which RSA is weak. This requires the
knowledge of a few most significant bits of (alternatively these bits need to be searched exhaustively). We provide experimental results to highlight that the problem can be solved with low lattice dimensions in practice. Our strategy outperforms the existing experimental results by increasing the bounds of . We provide clear evidence that RSA, with of the order of for 1000 bit , can be cryptanalysed in practice from the knowledge of .
Scratch, Click & Vote: E2E voting over the Internet
We present Scratch, Click & Vote voting scheme, which is a modification of the Punchscan and ThreeBallot systems. The scheme is end-to-end veryfiable and allows for voting over the Internet.
Security against malicious hardware and software used by a voter %TT
is due to the fact that a voter's computer does not get any knowledge about the voter's choice. Moreover, it can change successfully a voter's ballot only with a small probability.
As a side result, we present a modification of the ThreeBallot that eliminates Strauss'-like attacks on this scheme.
A new almost perfect nonlinear function which is not quadratic
We show how to change one coordinate function of an
almost perfect nonlinear
(APN) function in order to obtain new examples. It turns out that
this is a very powerful method to construct new
APN functions. In particular, we show that the approach can
be used to construct ``non-quadratic'' APN functions.
This new example
is in remarkable contrast to all recently constructed functions which
have all been quadratic.
Improved efficiency of Kiltz07-KEM
Uncategorized
Uncategorized
Kiltz proposed a practical key encapsulation mechanism(Kiltz07-KEM) which is secure
against adaptive chosen ciphertext attacks(IND-CCA2) under the gap hashed
Diffie-Hellman(GHDH) assumption\cite{Kiltz2007}. We show a variant of Kiltz07-KEM which
is more efficient than Kiltz07-KEM in encryption. The new scheme can be proved to be
IND-CCA2 secure under the same assumption, GHDH.
Treatment of the Initial Value in Time-Memory-Data Tradeoff Attacks on Stream Ciphers
Time-Memory Tradeoff (TMTO) attacks on stream ciphers are a
serious security threat and the resistance to this class of
attacks is an important criterion in the design of a modern stream
cipher. TMTO attacks are especially effective against stream
ciphers where a variant of the TMTO attack can make use of
multiple data to reduce the off-line and the on-line time
complexities of the attack (given a fixed amount of memory).
In this paper we present a new approach to TMTO attacks against
stream ciphers using a publicly known initial value (IV): We
suggest not to treat the IV as part of the secret key material (as
done in current attacks), but rather to choose in advance some IVs
and apply a TMTO attack to streams produced using these IVs. We
show that while the obtained tradeoff curve is identical to the
curve obtained by the current approach, the new technique allows
to mount the TMTO attack in a larger variety of settings. For
example, if both the secret key and the IV are of length n, it
is possible to mount an attack with data, time, and memory
complexities of 2^{4n/5}, while in the current approach, either
the time complexity or the memory complexity is not less than
2^n. We conclude that if the IV length of a stream cipher is
less than 1.5 times the key length, there exists an attack on the
cipher with data, time, and memory complexities less than the
complexity of exhaustive key search.
Attacks on RFID Protocols
This document consists of a collection of attacks upon RFID protocols and is meant to serve as a quick and easy reference. This document will be updated as new attacks are found. Currently the only attacks on protocols shown are the authors' original attacks with references to similar attacks on other
protocols.
The main security properties considered are authentication, untraceability, and - for stateful protocols - desynchronization resistance.
Revocation Systems with Very Small Private Keys
Uncategorized
Uncategorized
In this work, we design a method for creating public key broadcast
encryption systems. Our main technical innovation is based
on a new ``two equation'' technique for revoking users. This technique
results in two key contributions:
First, our new scheme has ciphertext size overhead ,
where is the number of revoked users, and the size of public and
private keys is only a \emph{constant} number of group elements from
an elliptic-curve group of prime order. In addition, the public key allows
us to encrypt to an unbounded number of users. Our system is the first to
achieve such parameters. We give two versions of our scheme: a simpler version which we prove to be selectively secure in the standard model under a new, but non-interactive assumption, and another version that employs the new dual system encryption technique of Waters to obtain adaptive security under the d-BDH and decisional Linear assumptions.
Second, we show that our techniques can be used to realize
Attribute-Based Encryption (ABE) systems with non-monotonic access
formulas, where our key storage is significantly more efficient than
previous solutions.
This result is also proven selectively secure in the standard model under our new non-interactive assumption.
We believe that our new technique will be of use elsewhere as well.
Strongly-Resilient and Non-Interactive Hierarchical Key-Agreement in MANETs
Key agreement is a fundamental security functionality by which pairs of nodes agree on shared keys to be used for protecting their pairwise communications. In this work we study key-agreement schemes that are well-suited for the mobile network environment.
Specifically, we describe schemes with the following haracteristics:
-- Non-interactive: any two nodes can compute a unique shared secret
key without interaction;
-- Identity-based: to compute the shared secret key, each node only
needs its own secret key and the identity of its peer;
-- Hierarchical: the scheme is decentralized through a
hierarchy where intermediate nodes in the hierarchy can derive the secret keys for each of its children without any limitations or prior knowledge on the number of such children or their identities;
-- Resilient: the scheme is fully resilient against compromise of
{\em any number of leaves} in the hierarchy, and of a threshold number of nodes in each of the upper levels of the hierarchy.
Several schemes in the literature have three of these four properties, but the schemes in this work are the first to possess all four. This makes them well-suited for environments such as MANETs and tactical networks which are very dynamic, have significant bandwidth and energy constraints, and where many nodes are vulnerable to compromise. We provide rigorous analysis of the proposed schemes and discuss implementations aspects.
Full Security:Fuzzy Identity Based Encryption
Uncategorized
Uncategorized
At EUROCRYPT 2005, Sahai and Waters presented the Fuzzy Identity Based Encryption (Fuzzy-IBE) which could be used for biometrics and attribute-based encryption in the selective-identity model. When a secure Fuzzy-IBE scheme in the selective-identity model is
transformed to full identity model it exist an exponential loss of security. In this paper, we use the CPA secure Gentry's IBE (exponent inversion IBE) to construct the first Fuzzy IBE that is fully secure without random oracles. In addition, the same technique is used to the modification of CCA secure Gentry's IBE which introduced by Kiltz and Vahlis to get the CCA secure Fuzzy IBE in the full-identity
model.
Combinatorial batch codes
In this paper, we study batch codes, which were
introduced by Ishai, Kushilevitz, Ostrovsky and Sahai.
A batch code specifies a method to distribute a
database of n items among m devices (servers)
in such a way that any k items
can be retrieved by reading at most t items from each of the servers. It is of interest to devise batch codes that
minimize the total storage, denoted by N, over all m servers.
In this paper, we study the special case t=1, under
the assumption that every server stores a subset of
the items. This is purely a combinatorial problem, so
we call this kind of batch code a "combinatorial batch code''.
For various parameter situations, we are able to present
batch codes that are optimal with respect to the storage
requirement, N. We also study uniform codes, where every item is
stored in precisely c of the m servers (such a code
is said to have rate 1/c). Interesting new results
are presented in the cases c = 2, k-2 and k-1. In addition,
we obtain improved existence results for arbitrary
fixed c using the probabilistic method.
Identity-Based Directed Signature Scheme from Bilinear Pairings
In a directed signature scheme, a verifier can exclusively verify
the signatures designated to himself, and shares with the signer the
ability to prove correctness of the signature to a third party when
necessary. Directed signature schemes are suitable for applications
such as bill of tax and bill of health.
This paper studies directed signatures in the identity-based setting. We first present the syntax and security notion that includes unforgeability and invisibility, then propose a concrete identity-based directed signature scheme from bilinear pairings. We then prove our scheme existentially unforgeable under the computational Diffie-Hellman
assumption, and invisible under the decisional Bilinear Diffie-Hellman assumption, both in the random oracle model.
A New Randomness Extraction Paradigm for Hybrid Encryption
We present a new approach to the design of IND-CCA2 secure hybrid encryption schemes in the standard model. Our approach provides an efficient generic transformation from 1-universal to 2-universal hash proof systems. The transformation involves a randomness extractor based on a 4-wise independent hash function as the key derivation function. Our methodology can be instantiated with efficient schemes based on standard intractability assumptions such as DDH, QR and Paillier. Interestingly, our framework also allows to prove IND-CCA2 security of a hybrid version of 1991's Damgaard's ElGamal public-key encryption
scheme under the DDH assumption.
Complete Fairness in Secure Two-Party Computation
In the setting of secure two-party computation, two mutually distrusting parties wish to compute some function of their inputs while preserving, to the extent possible, security properties such as privacy, correctness, and more. One desirable property is fairness which guarantees, informally, that if one party receives its output, then the other party does too. Cleve (STOC 1986) showed that complete fairness cannot be achieved, in general, without an honest majority. Since then, the accepted folklore has been that nothing non-trivial can be computed with complete fairness in the two-party setting, and the problem has been treated as closed since the late '80s.
In this paper, we demonstrate that this folklore belief is false by showing completely-fair protocols for various non-trivial functions in the two-party setting based on standard cryptographic assumptions. We first show feasibility of obtaining complete fairness when computing any function over polynomial-size domains that does not contain an ``embedded XOR''; this class of functions includes boolean AND/OR as
well as Yao's ``millionaires' problem''. We also demonstrate feasibility for certain functions that do contain an embedded XOR, and prove a lower bound showing that any completely-fair protocol for such functions must have round complexity super-logarithmic in the security parameter. Our results demonstrate that the question of completely-fair secure computation without an honest majority is far from closed.
Secure Biometric Authentication With Improved Accuracy
We propose a new hybrid protocol for cryptographically secure biometric authentication. The main advantages of the proposed protocol over previous solutions can be summarised as follows: (1) potential for much better accuracy using different types of biometric signals, including behavioural ones; and (2) improved user privacy, since user identities are not transmitted at any point in the protocol execution. The new protocol takes advantage of state-of-the-art identification classifiers, which provide not only better accuracy, but also the possibility to perform authentication without knowing who the user claims to be. Cryptographic security is based on the Paillier public key encryption
scheme.
Accountability of Perfect Concurrent Signature
Concurrent signature provided a novel idea for fair exchange protocol without trusted third party. Perfect Concurrent Signature is proposed to strengthen theambiguity of the concurrent signature. Wang et al, pointed out there exist an attack against the fairness of Perfect Concurrent Signature and proposed the improved perfect
concurrent signature. This paper find that in proposed (perfect) concurrent signature protocol, no matter two party or multi-party, the signer could bind multiple messages with one keystone set but let the other signers know only one of the messages. This is a
new unfair case in the application of concurrent signature. Based on this observation, we propose that accountability should be one of the security properties of (perfect) concurrent signature and we give the definition of accountability of concurrent signature. To illustrate this idea, we give an attack scene against the accountability of
improved perfect concurrent signature proposed by Wang et al, and propose an update
version of perfect concurrent signature to avoid such attack.
Cheon's algorithm, pairing inversion and the discrete logarithm problem
We relate the fixed argument pairing inversion problems (FAPI) and the discrete logarithm problem on an elliptic curve. This is done using the reduction from the DLP to the Diffie-Hellman problem developed by Boneh, Lipton, Maurer and Wolf. This approach fails when only one of the FAPI problems can be solved. In this case we use Cheon's algorithm to get a reduction.
An analysis of the infrastructure in real function fields
We construct a map injecting the set of infrastructure ideals in a real function field into the class group of the correspoding hyperelliptic curve. This map respects the `group-like' structure of the infrastructure, as a consequence of this construction we show that calculating distances in the set of infrastructure ideals is equivalent to the DLP in the underlying hyperelliptic curve. We also give a precise description of the elements missing in the infrastructure to be a group.
Nonlinear Piece In Hand Perturbation Vector Method for Enhancing Security of Multivariate Public Key Cryptosystems
Abstract. The piece in hand (PH) is a general scheme which is applicable to any reasonable type of multivariate public key cryptosystems for the purpose of enhancing their security. In this paper, we propose a new class PH method called NLPHPV (NonLinear Piece in Hand Perturbation Vector) method. Although our NLPHPV uses
similar perturbation vectors as is used for the previously known internal perturbation method, this new method can avoid redundant repetitions in decryption process. With properly chosen parameter sizes, NLPHPV achieves an observable gain in security from the original multivariate public key cryptosystem. We demonstrate these by both theoretical analyses and computer simulations against major known attacks and provides the concrete sizes of security parameters, with which we even expect the grater security against potential quantum attacks.
Attack on Kang et al.'s Identity-Based Strong Designated Verifier Signature Scheme
Uncategorized
Uncategorized
In this paper, we present a universal forgery attack on Kang et al.'s identity-based strong designated verifier signature (IBSDVS) scheme. We show anyone can forge a valid IBSDVS on an arbitrary message without the knowledge of the private key of either the signer or the designated verifier. Moreover, we point out that Kang et al.'s scheme does not satisfy the properties of strongness and non-delegatability. At last, an improved IBSDVS scheme for Kang et al.'s scheme is presented, and it is provably secure and achieves all the requirements for an IBSDVS.
Cryptanalysis of Short Exponent RSA with Primes Sharing Least Significant Bits
LSBS-RSA denotes an RSA system with modulus primes, p and q, sharing a large number of least significant bits. In ISC 2007, Zhao and Qi analyzed the security of short exponent LSBS-RSA. They claimed that short exponent LSBS-RSA is much more vulnerable to the lattice attack than the standard RSA. In this paper, we point out that there exist some errors in the calculation of Zhao & Qi's attack. After re-calculating, the result shows that their attack is unable for attacking RSA with primes sharing bits. Consequently, we give a revised version to make their attack feasible. We also propose a new method to further extend the security boundary, compared with the revised version. The proposed attack also supports the result of analogue Fermat factoring on LSBS-RSA, which claims that p and q cannot share more than (n/4) least significant bits, where n is the bit-length of pq. In conclusion, it is a trade-off between the number of sharing bits and the security level in LSBS-RSA. One should be more careful when using LSBS-RSA with short exponents.
Foundations of Group Key Management – Framework, Security Model and a Generic Construction
Group Key Establishment is fundamental for a variety of security mechanisms in group applications. It allows n > 1 principals to agree upon a common secret key. This can further be classified into Group Key Exchange (or Group Key Agreement), where all the principals participate in the construction of the key, and Group Key Transport (or Group Key Distribution), where the key is chosen by a singe principal and is then securely communicated to the others. Both these techniques can be analyzed in the context of either static or dynamic groups. Dynamic Group Key Establishment is better known as Group Key Management (GKM), as it involves not only the initital key establishment, but also efficient key management when group members join or leave the group. Dynamic Group Key Exchange is also known as decentralized or distributed GKM, while Dynamic Group Key Transport is known as centralized GKM. While there has been a lot of recent work in formal security models for Dynamic Group Key Exchange, little, if any, attention has been directed towards building a concrete framework and formal security model for centralized GKM. Many such schemes that have been proposed so far have been broken, as they cite ambiguous arguments and lack formal proofs. In this paper, we take a first step towards addressing this problem by providing firm foundations for centralized Group Key Management. We provide a generalized framework for centralized GKM along with a formal security model and strong definitions for the security properties that dynamic groups demand. We also show a generic construction of a centralized GKM scheme from any given multi-receiver ID-based Key Encapsulation Mechanism (mID-KEM). By doing so, we unify two concepts that are significantly different in terms of what they achieve. Our construction is simple and efficient. We prove that the resulting GKM inherits the security of the underlying mID-KEM up to CCA security. We also illustrate our general conversion using the mID-KEM proposed in 2007 by Delerablée.
A New Message Recognition Protocol for Ad Hoc Pervasive Networks
We propose a message recognition protocol which is suitable for ad hoc pervasive networks without the use of hash chains. Hence, we no longer require the sensor motes to save values of a hash chain in their memories. This relaxes the memory requirements. Moreover, we do not need to fix the total number of times the protocol can be executed which implies a desired flexibility in this regard. Furthermore, our protocol is secure without having to consider families of assumptions that depend on the number of sessions the protocol is executed. Hence, the security does not weaken as the protocol is executed over time. Last but not least, we provide a practical procedure for resynchronization in case of any adversarial disruption or communication failure.
Maximizing data survival in Unattended Wireless Sensor Networks against a focused mobile adversary
Some sensor network settings involve disconnected or
unattended operation with periodic visits by a mobile sink.
An unattended sensor network operating in a hostile environment can
collect data that represents a high-value target for the adversary.
Since an unattended sensor can not immediately off-load sensed data to
a safe external entity (such as a sink), the adversary can easily
mount a focused attack aiming to erase or modify target data.
To maximize chances of data survival, sensors must collaboratively attempt to mislead the adversary and hide the location, the origin and the contents of collected data.
In this paper, we focus on applications of well-known security
techniques to maximize chances of data survival in unattended sensor
networks, where sensed data can not be off-loaded to a sink in real time.
Our investigation yields some interesting insights and surprising results.
The highlights of our work are:
(1) thorough exploration of the data survival challenge,
(2) exploration of the design space for possible solutions,
(3) construction of several practical and effective techniques,
and (4) their evaluation.
Another approach to pairing computation in Edwards coordinates
Uncategorized
Uncategorized
The recent introduction of Edwards curves has significantly reduced
the cost of addition on elliptic curves. This paper presents new
explicit formulae for pairing implementation in Edwards coordinates.
We prove our method gives performances similar to those of Miller's
algorithm in Jacobian coordinates and is thus of cryptographic
interest when one chooses Edwards curve implementations of protocols
in elliptic curve cryptography. The method is faster than the recent
proposal of Das and Sarkar for computing pairings on supersingular
curves using Edwards coordinates.
How to Protect Yourself without Perfect Shredding
Erasing old data and keys is an important tool in cryptographic protocol design. It is useful in many settings, including proactive security, adaptive security, forward security, and intrusion resilience. Protocols for all these settings typically assume the ability to perfectly erase information. Unfortunately, as amply demonstrated in the systems literature, perfect erasures are hard to implement in practice.
We propose a model of partial erasures where erasure instructions leave almost all the data erased intact, thus giving the honest players only a limited capability for disposing of old data. Nonetheless, we provide a general compiler that transforms any secure protocol using perfect erasures into one that maintains the same security properties when only partial erasures are available. The key idea is a new redundant representation of secret data which can still be computed on, and yet is rendered useless when partially erased. We prove that any such a compiler must incur a cost in additional storage, and that our compiler is near optimal in terms of its storage overhead.
Ciphertext-Policy Attribute-Based Encryption: An Expressive, Efficient, and Provably Secure Realization
Uncategorized
Uncategorized
We present a new methodology for realizing Ciphertext-Policy Attribute
Encryption (CP-ABE) under concrete and noninteractive cryptographic
assumptions in the standard model. Our solutions allow any encryptor
to specify access control in terms of any access formula over the
attributes in the system. In our most efficient system, ciphertext size, encryption,
and decryption time scales linearly with the complexity of the access
formula. The only previous work to achieve these parameters was
limited to a proof in the generic group model.
We present three constructions within our framework.
Our first system is proven selectively secure under a assumption that
we call the decisional Parallel Bilinear Diffie-Hellman Exponent
(PBDHE) assumption which can be viewed as a generalization of the BDHE
assumption. Our next two constructions
provide performance tradeoffs to achieve provable security
respectively under the (weaker) decisional Bilinear-Diffie-Hellman
Exponent and decisional Bilinear Diffie-Hellman assumptions.
Sharemind: a framework for fast privacy-preserving computations
Gathering and processing sensitive data is a difficult task. In fact, there is no common recipe for building the necessary information systems. In this paper, we present a provably secure and efficient general-purpose computation system to address this problem. Our solution - SHAREMIND - is a virtual machine for privacy-preserving data processing that relies on share computing techniques. This is a standard way for securely evaluating functions in a multi-party computation environment. The novelty of our solution is in the choice of the secret sharing scheme and the design of the protocol suite. We have made many practical decisions to make large-scale share computing feasible in practice. The protocols of SHAREMIND are information-theoretically secure in the honest-but-curious model with three computing participants. Although the honest-but-curious model does not tolerate malicious participants, it still provides significantly increased privacy preservation when compared to standard centralised databases.
How to Launch A Birthday Attack Against DES
We present a birthday attack against DES. It is entirely based on the relationship and the simple key schedule in DES.
It requires about ciphertexts of the same , encrypted by the same key . We conjecture it has a computational complexity of . Since the requirement for the birthday attack is more accessible than that for Differential cryptanalysis, Linear cryptanalysis or Davies' attack, it is of more practical
significance.
Authenticated Byzantine Generals in Dual Failure Model
Uncategorized
Uncategorized
Pease {\em et al.}\/ introduced the problem of Byzantine Generals (BGP) to study the effects of Byzantine faults in distributed protocols for reliable broadcast. It is well known that BGP among players tolerating up to faults is (efficiently) possible if and only if . To overcome this severe limitation, Pease {\em et al.} introduced a variant of BGP, \emph{Authenticated Byzantine General} (ABG). Here players are supplemented with digital signatures (or similar tools) to thwart the challenge posed by Byzantine faults. Subsequently, they proved that with the use of authentication, fault tolerance of protocols for reliable broadcast can be amazingly increased to (which is a huge improvement over the ).
Byzantine faults are the most generic form of faults. In a network not {\em all} faults are always malicious. Some faulty nodes may only leak their data while others are malicious. Motivated from this, we study the problem of ABG in ( , )-mixed adversary model where the adversary can corrupt up to any players actively and control up to any other players passively. We prove that in such a setting, ABG over a completely connected synchronous network of nodes tolerating a ( , )-adversary is possible if and only if +min( ) when . Interestingly, our results can also be seen as an attempt to unify the extant literature on BGP and ABG.
One-Up Problem for (EC)DSA
The one-up problem is defined for any group and a function from the group to integers. An instance of problem consists of two points in the group. A solution consists of another point satisfying an equation involving the group and the function. The one-up problem must be hard for the security of (EC)DSA. Heuristic arguments are given for the independence of the discrete log and one-up problems. DSA and ECDSA are conjectured to be secure if the discrete log and one-up problems are hard and the hash is collision resistant.
Hybrid Binary-Ternary Joint Sparse Form and its Application in Elliptic Curve Cryptography
Uncategorized
Uncategorized
Multi-exponentiation is a common and time consuming operation in public-key cryptography. Its elliptic curve counterpart, called multi-scalar multiplication is extensively used for digital signature verification. Several algorithms have been proposed to speed-up those critical computations. They are based on simultaneously recoding a set of integers in order to minimize the number of general multiplications or point additions. When signed-digit recoding techniques can be used, as in the world of elliptic curves, Joint Sparse Form (JSF) and interleaving -NAF are the most efficient algorithms. In this paper, a novel recoding algorithm for a pair of integers is proposed, based on a decomposition that mixes powers of 2 and powers of 3. The so-called Hybrid Binary-Ternary Joint Sparse Form require fewer digits and is sparser than the JSF and the interleaving -NAF. Its advantages are illustrated for elliptic curve double-scalar multiplication; the operation counts show a gain of up to 18\%.
Breaking the Akiyama-Goto cryptosystem
Akiyama and Goto have proposed a cryptosystem based on rational points on curves over function fields (stated in the equivalent form of sections of fibrations on surfaces). It is easy to construct a curve passing through a few given points, but finding the points, given only the curve, is hard. We show how to break their original cryptosystem by using algebraic points instead of rational points and discuss possibilities for changing their original system to create a secure one.
Attacks on Singelee and Preneel's protocol
Singelee and Preneel have recently proposed a enhancement of
Hancke and Kuhn's distance bounding protocol for RFID. The authors
claim that their protocol offers substantial reductions in the
number of rounds, though preserving its advantages: suitable to be
employed in noisy wireless environments, and requiring so few
resources to run that it can be implemented on a low-cost device.
Subsequently, the same authors have also proposed it as an
efficient key establishment protocol in wireless personal area
networks. Nevertheless, in this paper we show effective relay
attacks on this protocol, which dramatically increase the success
probability of an adversary. As a result, the effectiveness of
Singelee and Preneel's protocol is seriously questioned.
Survival in the Wild: Robust Group Key Agreement in Wide-Area Networks
Group key agreement (GKA) allows a set of players to establish a shared secret and thus bootstrap secure group communication. GKA is very useful in many types of peer group scenarios and applications. Since all GKA protocols involve multiple rounds, robustness to player failures is important and desirable. A robust group key agreement (RGKA) protocol runs to completion even if some players fail during protocol execution.
Previous work yielded constant-round RGKA protocols suitable for the LAN setting, assuming players are homogeneous, failure probability is uniform and player failures are independent. However, in a more general widearea network (WAN) environment, heterogeneous hardware/software and communication facilities can cause wide variations in failure probability among players. Moreover, congestion and communication equipment failures can result in correlated failures among subsets of GKA players.
In this paper, we construct the first RGKA protocol that supports players with different failure probabilities, spread across any LAN/WAN combination, while also allowing for correlated failures among subgroups of players. The proposed protocol is efficient (2 rounds) and provably secure. We evaluate its robustness and performance both analytically and via simulations.
Linear and Differential Cryptanalysis of Reduced SMS4 Block Cipher
SMS4 is a 128-bit block cipher with a 128-bit user key and 32 rounds, which is used in WAPI, the Chinese WLAN national standard. In this paper, we present a linear attack and a differential attack on a 22-round reduced SMS4; our 22-round linear attack has a data complexity of 2^{117} known plaintexts, a memory complexity of 2^{109} bytes and a time complexity of 2^{109.86} 22-round SMS4 encryptions and 2^{120.39} arithmetic operations, while our 22-round differential attack requires 2^{118} chosen plaintexts, 2^{123} memory bytes and 2^{125.71} 22-round SMS4 encryptions. Both of our attacks are better than any previously known cryptanalytic results on SMS4 in terms of the number of attacked rounds. Furthermore, we present a boomerang and a rectangle attacks on a 18-round reduced SMS4. These results are better than previously known rectangle attacks on reduced SMS4. The methods presented to attack SMS4 can be applied to other unbalanced Feistel ciphers with incomplete diffusion.
FPGA and ASIC Implementations of the Pairing in Characteristic Three
Since their introduction in constructive cryptographic applications, pairings over (hyper)elliptic curves are at the heart of an ever increasing number of protocols. As they rely critically on efficient algorithms and implementations of pairing primitives, the study of hardware accelerators became an active research area.
In this paper, we propose two coprocessors for the reduced pairing introduced by Barreto {\it et al.} as an alternative means of computing the Tate pairing on supersingular elliptic curves. We prototyped our architectures on FPGAs. According to our place-and-route results, our coprocessors compare favorably with other solutions described in the open literature. We also present the first ASIC implementation of the reduced pairing.
Delegating Capabilities in Predicate Encryption Systems
In predicate encryption systems, given a capability, one can evaluate one or more predicates on the encrypted data, while all other information about the plaintext remains hidden. We consider the first such systems to permit delegation of capabilities. In a system that supports delegation, a user Alice who has a capability can delegate to Bob a more restrictive capability, which allows him to learn less about encrypted data than she did.
We formally define delegation in predicate encryption systems, and propose a new security definition for delegation. In addition, we present an efficient construction supporting conjunctive queries. The security of our construction can be reduced to the general 3-party Bilinear Diffie-Hellman assumption, and the Bilinear Decisional Diffie-Hellman assumption in composite-order bilinear groups.
An Improved Robust Fuzzy Extractor
We consider the problem of building robust fuzzy extractors, which allow two parties holding similar random variables W, W' to agree on a secret key R in the presence of an active adversary. Robust fuzzy extractors were defined by Dodis et al. in Crypto 2006 to be noninteractive, i.e., only one message P, which can be modified by an unbounded adversary, can pass from one party to the other. This allows them to be used by a single party at different points in time (e.g., for key recovery or biometric authentication), but also presents an additional challenge: what if R is used, and thus possibly observed by the adversary, before the adversary has a chance to modify P. Fuzzy extractors secure against such a strong attack are called post-application robust.
We construct a fuzzy extractor with post-application robustness that extracts a shared secret key of up to (2m-n)/2 bits (depending on error-tolerance and security parameters), where n is the bit-length and m is the entropy of W. The previously best known result, also of Dodis et al., extracted up to (2m-n)/3 bits (depending on the same parameters).
A strategy for any DAA Issuer and an additional verification by a Host
A successful strategy was identified
for any Verifier colluding with any Issuer
to distinguish honest Provers issuing DAA signatures.
An additional verification equation was introduced
for a Prover to detect 'tagged' credentials
that may be issued while Join protocol.
This verification can be done by the Host
and do not affect TPM in any way.
Signcryption with Proxy Re-encryption
Confidentiality and authenticity are two of the most fundamental problems in cryptography. Many applications require both confidentiality and authenticity, and hence an efficient way to get both together was very desirable. In 1997, Zheng proposed the notion of ``signcryption'', a single primitive which provides both confidentiality and authenticity in a way that's more efficient than signing and encrypting separately. Proxy re-encryption is a primitive that allows a semi-trusted entity called the ``proxy'' to convert ciphertexts addressed to a ``delegator'' to those that can be decrypted by a ``delegatee'', by using some special information given by the delegator, called the ``rekey''. In this work, we propose the notion of signcryption with proxy re-encryption (SCPRE), and motivate the same. We define security models for SCPRE, and also propose a concrete unidirectional, non-interactive identity-based SCPRE construction. We also provide complete proofs of security for the scheme in the security models defined. We finally provide directions for further research in this area.
Certificate-Based Signature Schemes without Pairings or Random Oracles
In this paper, we propose two new certificate-based signature
(CBS) schemes with new features and advantages. The first one
is very efficient as it does not require any pairing
computation and its security can be proven using Discrete
Logarithm assumption in the random oracle model. We also
propose another scheme whose security can be proven in the
standard model without random oracles. To the best of our
knowledge, these are the \emph{first} CBS schemes in the
literature that have such kind of features.
Twisted Ate Pairing on Hyperelliptic Curves and Applications
Uncategorized
Uncategorized
In this paper we show that the twisted Ate pairing on elliptic curves can be generalized to hyperelliptic curves, we also give a series of variations of the hyperelliptic Ate and twisted Ate pairings. Using the hyperelliptic Ate pairing and twisted Ate
pairing, we propose a new approach to speed up the Weil pairing computation, and obtain an interested result: For some hyperelliptic curves with high degree twist, using this approach to compute Weil pairing will be faster than Tate pairing, Ate pairing etc. all known pairings.
White-Box Cryptography: Formal Notions and (Im)possibility Results
A key research question in computer security is whether
one can implement software that offers some protection
against software attacks from its execution platform. While
code obfuscation attempts to hide certain characteristics of
a program P, white-box cryptography specifically focusses
on software implementations of cryptographic primitives
(such as encryption schemes); the goal of a white-box implementation
is to offer a certain level of robustness against
an adversary who has full access to and control over the
implementation of the primitive. Several formal models for
obfuscation have been presented before, but it is not clear if
any of these definitions can capture the concept of white-box
cryptography. In this paper, we discuss the relation between
obfuscation and white-box cryptography, and formalize the
notion of white-box cryptography by capturing the security
requirement using a 'White-Box Property' (WBP). In
the second part, we present positive and negative results on
white-box cryptography. We show that for interesting programs
(such as encryption schemes, and digital signature
schemes), there are security notions that cannot be satisfied
when adversaries have white-box access, while the notion
is satisfied when the adversary has black-box access to its
functionality. On the positive side, we show that there exists
an obfuscator for a symmetric encryption scheme for which
a useful security notion (such as CPA security) remains satisfied
when an adversary has access to its white-box implementation.
A New Hash Family Obtained by Modifying the SHA-2 Family
Uncategorized
Uncategorized
In this work, we study several properties of the SHA-2 design which have been utilized in recent collision attacks against reduced round SHA-2. Small modifications to the SHA-2 design are suggested to thwart these attacks. The modified round function provides the
same resistance to linearization attacks as the original SHA-2 round function, but, provides better resistance to non-linear attacks. Our next contribution is to introduce the general idea of ``multiple feed-forward" for the construction of cryptographic hash functions.
This can provide increased resistance to the Chabaud-Joux type ``perturbation-correction'' collision attacks. The idea of feed-forward is taken further by introducing the idea of feed-forward across message blocks leading to resistance against generic multi-collision attacks. The net effect of the suggested changes to the SHA-2 design has insignificant impact on the efficiency of computing the digest.
A Combinatorial Analysis of Recent Attacks on Step Reduced SHA-2 Family
Uncategorized
Uncategorized
We perform a combinatorial analysis of SHA-2 compression function. This analysis explains in a unified way the recent attacks against reduced round SHA-2. We start with a general class of local collisions and show that the previously used local collision by
Nikolić and Biryukov (NB) and Sanadhya and Sarkar (SS) are special cases. The study also clarifies several advantages of the SS local collision over the NB local collision. Deterministic constructions of up to 22-round SHA-2 collisions are described
using the SS local collision and up to 21-round SHA-2 collisions are described using the NB local collision. For 23 and 24-round SHA-2, we describe a general strategy and then apply the SS local collision to this strategy. The resulting attacks are faster than those proposed by Indesteege et al using the NB local collision. We provide colliding message pairs for 22, 23 and 24-round SHA-2.
Although these attacks improve upon the existing reduced round
SHA-256 attacks, they do not threaten the security of the full SHA-2 family.
\footnote{This work builds upon and subsumes previous work done by us. Whereas the previous works focused on obtaining collisions for fixed number of rounds, the current work provides the combinatorial framework for understanding how such collisions arise.}
New Collision attacks Against Up To 24-step SHA-2
Uncategorized
Uncategorized
In this work, we provide new and improved attacks against 22, 23 and 24-step SHA-2 family using a local collision given by Sanadhya and Sarkar (SS) at ACISP '08. The success probability of our 22-step attack is 1 for both SHA-256 and SHA-512.
The computational efforts for the 23-step and 24-step SHA-256 attacks
are respectively and calls to the corresponding step reduced SHA-256. The corresponding values for the 23 and 24-step SHA-512 attack are respectively and calls.
Using a look-up table having (resp. ) entries the computational effort for finding 24-step SHA-256 (resp. SHA-512)
collisions can be reduced to (resp. ) calls.
We exhibit colliding message pairs for 22, 23 and 24-step SHA-256 and SHA-512. This is the \emph{first} time that a colliding message pair for 24-step SHA-512 is provided.
The previous work on 23 and 24-step SHA-2 attacks is due to Indesteege et al. and utilizes the local collision presented by Nikolić and Biryukov NB) at FSE '08. The reported computational efforts are and for 23 and 24-step SHA-256 respectively and and for 23 and 24-step SHA-512. The previous 23 and 24-step attacks first constructed a pseudo-collision and later converted it into a collision for the reduced round SHA-2 family. We show that this two step procedure is unnecessary. Although these attacks improve upon the existing reduced round SHA-2 attacks, they do not threaten the security of the full SHA-2 family.
Searching for Low Weight Codewords in Linear Binary Codes
Uncategorized
Uncategorized
In this work we revisit the known algorithms for searching for low weight codewords in linear binary codes. We propose some improvements on them and also propose a new efficient heuristic.
Adaptive Security in Broadcast Encryption Systems
Uncategorized
Uncategorized
We present new techniques for achieving adaptive security
in broadcast encryption systems. Previous work on fully-collusion
resistant broadcast encryption with short ciphertexts was limited
to only considering static security.
First, we present a new definition of security that we call
semi-static security and show a generic ``two-key"
transformation from semi-statically secure systems to adaptively
secure ones that have comparable-sized ciphertexts. Using bilinear
maps, we then construct broadcast encryption systems that are
semi-statically secure in the standard model and have constant
size ciphertexts. Our semi-static constructions work when the
number of indices or identifiers in the system is polynomial in
the security parameter.
For identity-based broadcast encryption, where the number of
potential indices or identifiers may be exponential, we present
the first adaptively secure system with sublinear ciphertexts. We
prove security in the standard model.
Deterministic Encryption: Definitional Equivalences and Constructions without Random Oracles
We strengthen the foundations of deterministic public-key encryption via definitional equivalences and standard-model constructs based on general assumptions. Specifically we consider seven notions of privacy for deterministic encryption, including six forms of semantic security and an indistinguishability notion, and show them all equivalent. We then present a deterministic scheme for the secure encryption of uniformly and independently distributed messages based solely on the existence of
trapdoor one-way permutations. We show a generalization of the construction that allows secure deterministic encryption of independent high-entropy messages. Finally we show relations between deterministic and standard (randomized) encryption.
Information-Theoretically Secure Voting Without an Honest Majority
We present three voting protocols with unconditional privacy and information-theoretic correctness, without assuming any bound on the
number of corrupt voters or voting authorities. All protocols have
polynomial complexity and require private channels and a simultaneous broadcast channel. Our first protocol is a basic voting scheme which allows voters to interact in order to compute the tally. Privacy of the ballot is unconditional, but any voter can cause the protocol to fail, in which case information about the tally may nevertheless transpire. Our second protocol introduces voting authorities which allow the implementation of the first protocol, while reducing the interaction and limiting it to be only between voters and authorities and among the authorities themselves. The simultaneous broadcast is also limited to the authorities. As long as a single authority is honest, the privacy is unconditional, however, a single corrupt authority or a single corrupt voter can cause the protocol to fail. Our final protocol provides a safeguard against corrupt voters by enabling a verification technique to allow the authorities to revoke incorrect votes. We also discuss the implementation of a simultaneous broadcast channel with the use of temporary computational assumptions, yielding versions of our protocols achieving everlasting security.
Efficient Hyperelliptic Arithmetic using Balanced Representation for Divisors
We discuss arithmetic in the Jacobian of a
hyperelliptic curve of genus . The traditional approach is to
fix a point and represent divisor classes in the
form where is effective and .
We propose a different representation which is balanced at infinity.
The resulting arithmetic is more efficient than previous approaches
when there are 2 points at infinity.
This is a corrected and extended version of the article presented in ANTS 2008. We include an appendix with explicit formulae to compute a very efficient `baby step' in genus 2 hyperelliptic curves given by an imaginary model.
Secure Computability of Functions in the IT setting with Dishonest Majority and Applications to Long-Term Security
Uncategorized
Uncategorized
It is well known that general secure function evaluation (SFE) with information-theoretical (IT) security is infeasible in presence of a corrupted majority in the standard model. On the other hand, there are SFE protocols (Goldreich et al.~[STOC'87]) that are computationally secure (without fairness) in presence of an actively corrupted majority of the participants. Now, the issue with computational assumptions is not so much that they might be unjustified at the time of protocol execution. Rather, we are usually worried about a potential violation of the privacy of sensitive data by an attacker whose power increases over time (e.g.~due to new technical developments). Therefore, we ask which functions can be computed with long-term security, where we admit computational assumptions for the duration of a computation, but require IT security (privacy) once the computation is concluded.
Toward this end we combinatorially characterize the classes of functions that can be computed IT securely in the authenticated channels model in presence of passive, semi-honest, active, and quantum adversaries (our results for quantum adversaries and in part for active adversaries are limited to the 2-party setting). In particular we obtain results on the fair computability of functions in the IT setting along the lines of the work of Gordon et al.~[STOC'08] for the computational setting. Our treatment is constructive in the sense that if a function is computable in a given setting, then we exhibit a protocol.
We show that the class of functions computable with long-term security in a very practical setting where the adversary may be active and insecure channels and a public-key infrastructure are provided is precisely the class of functions computable with IT security in the authenticated channels model in presence of a semi-honest adversary.
Finally, from our results and the work of Kushilevitz [SIAM Journal on Discrete Mathematics '92] and Kraschewski and Müller-Quade we can derive a complete combinatorial classification of functions, by secure computability and completeness under passive, semi-honest, active, and quantum adversaries.
Slide Attacks on a Class of Hash Functions
This paper studies the application of slide attacks to hash functions.
Slide attacks have mostly been used for block cipher cryptanalysis. But, as
shown in the current paper, they also form a potential threat for hash
functions, namely for sponge-function like structures. As it turns
out, certain constructions for hash-function-based MACs can be vulnerable to
forgery and even to key recovery attacks. In other cases, we can at least
distinguish a given hash function from a random oracle.
To illustrate our results, we describe attacks against the Grindahl-256 and
Grindahl-512 hash functions. To the best of our knowledge, this is the first
cryptanalytic result on Grindahl-512. Furthermore, we point out a
slide-based distinguisher attack on a slightly modified version of
RadioGatun. We finally discuss simple countermeasures as a defense against
slide attacks.
Statistically Reliable and Secure Message Transmission in Directed Networks
Consider the following problem: a sender S and a receiver R are part of a directed synchronous network and connected through intermediate nodes. Specifically, there exists n node disjoint paths, also called as wires, which are directed from S to R and u wires, which are directed from R to S. Moreover, the wires from S to R are disjoint from the wires directed from R to S. There exists a centralized, static adversary who has unbounded computing power and who can control at most t wires between S and R in Byzantine fashion. S has a message m^S, which we wants to send to R. The challenge is to design a protocol, such that after interacting in phases as per the protocol,
R should correctly output m^R = m^S, except with error probability
2^{-\Omega(\kappa)}, where \kappa is the error parameter. This problem is called as statistically reliable message transmission (SRMT). The problem of statistically secure message transmission
(SSMT) has an additional requirement that at the end of the protocol, m^S should be information theoretically secure.
Desmedt et.al have given the necessary and sufficient condition for the existence of SRMT and SSMT protocols in the above settings. They also presented an SSMT protocol, satisfying their characterization. Desmedt et.al claimed that their protocol is efficient and has polynomial computational and communication complexity. However, we show that it is not so. That is, we specify an adversary strategy, which may cause the protocol to have exponential computational and communication complexity. We then present new and efficient SRMT and SSMT protocols, satisfying the characterization of Desmedt et.al Finally we show that the our proposed protocols are communication optimal by deriving lower bound on the communication complexity of SRMT and SSMT protocols. As far our knowledge is concerned, our protocols are the first communication optimal SRMT and SSMT protocols in directed networks.
The Hidden Root Problem
In this paper we study a novel computational problem called the Hidden Root Problem, which appears naturally when considering fault attacks on pairing based cryptosystems. Furthermore, a variant of this problem is one of the main obstacles for efficient pairing inversion. We present an algorithm to solve this problem over extension fields and investigate for which parameters the algorithm becomes practical.
Breaking RSA Generically is Equivalent to Factoring
Uncategorized
Uncategorized
We show that a generic ring algorithm for breaking RSA in
can be converted into an algorithm for factoring the
corresponding RSA-modulus . Our results imply that any attempt at
breaking RSA without factoring will be non-generic and hence
will have to manipulate the particular bit-representation of the
input in . This provides new evidence that breaking RSA may be equivalent to factoring the modulus.
2-Adic Complexity of a Sequence Obtained from a Periodic Binary Sequence by Either Inserting or Deleting k Symbols within One Period
In this paper, we propose a method to get the lower bounds of the 2-adic complexity of a sequence obtained from a periodic sequence over GF(2) by either inserting or deleting k symbols within one period. The results show the variation of the distribution of the 2-adic complexity becomes as k increases. Particularly, we discuss the lower bounds when k respectively.
ON A CRYPTOGRAPHIC IDENTITY IN OSBORN LOOPS
Uncategorized
Uncategorized
This study digs out some new algebraic properties of an Osborn loop
that will help in the future to unveil the mystery behind the middle
inner mappings of an Osborn loop. These new algebraic
properties, will open our eyes more to the study of Osborn loops
like CC-loops which has received a tremendious attention in this
and VD-loops whose study is yet to be explored. In
this study, some algebraic properties of non-WIP Osborn loops have
been investigated in a broad manner. Huthnance was able to deduce
some algebraic properties of Osborn loops with the WIP i.e universal
weak WIPLs. So this work exempts the WIP. Two new loop identities,
namely left self inverse property loop(LSIPL) identity and right
self inverse property loop(RSLPL) are introduced for the first time
and it is shown that in an Osborn loop, they are equivalent. A
CC-loop is shown to be power associative if and only if it is a
RSLPL or LSIPL. Among the few identities that have been established
for Osborn loops, one of them is recognized and recommended for
cryptography in a similar spirit in which the cross inverse property
has been used by Keedwell following the fact that it was observed
that Osborn loops that do not have the LSIP or RSIP or 3-PAPL or
weaker forms of inverse property, power associativity and
diassociativity to mention a few, will have cycles(even long ones).
These identity is called an Osborn cryptographic identity(or just a
cryptographic identity).
ON MIDDLE UNIVERSAL -INVERSE QUASIGROUPS AND THEIR APPLICATIONS TO CRYPTOGRAPHY
Uncategorized
Uncategorized
This study presents a special type of middle isotopism under which
-inverse quasigroups are isotopic invariant. A sufficient
condition for an -inverse quasigroup that is specially isotopic
to a quasigroup to be isomorphic to the quasigroup isotope is
established. It is shown that under this special type of middle
isotopism, if is a positive even integer, then, a quasigroup is
an -inverse quasigroup with an inverse cycle of length if
and only if its quasigroup isotope is an -inverse quasigroup with
an inverse cycle of length . But when is an odd positive
integer. Then, if a quasigroup is an -inverse quasigroup with an
inverse cycle of length , its quasigroup isotope is an
-inverse quasigroup with an inverse cycle of length if and
only if the two quasigroups are isomorphic. Hence, they are
isomorphic -inverse quasigroups. Explanations and procedures are
given on how these results can be used to apply -inverse
quasigroups to cryptography, double cryptography and triple
cryptography.
ON MIDDLE UNIVERSAL WEAK AND CROSS INVERSE PROPERTY LOOPS WITH EQUAL LENGHT OF INVERES CYCLES
Uncategorized
Uncategorized
This study presents a special type of middle isotopism under which
the weak inverse property(WIP) is isotopic invariant in loops. A
sufficient condition for a WIPL that is specially isotopic to a loop
to be isomorphic to the loop isotope is established. Cross inverse
property loops(CIPLs) need not satisfy this sufficient condition. It
is shown that under this special type of middle isotopism, if is
a positive even integer, then a WIPL has an inverse cycle of length
if and only if its isotope is a WIPL with an inverse cycle of
length . But, when is an odd positive integer. If a loop or
its isotope is a WIPL with only and inverse cycles of length
, its isotope or the loop is a WIPL with only and inverse
cycles of length if and only if they are isomorphic. So, that
both are isomorphic CIPLs. Explanations and procedures are given on
how these results can be used to apply CIPLs to cryptography.
Embedding in Two Least Significant Bits with Wet Paper Coding
In this paper, we present three embedding schemes for extensions of least significant bit overwriting to both of the two lowest bit planes in digital images. Our approaches are inspired by the work of Fridrich et al. [8] who proposed wet paper coding as an efficient method for steganographic schemes. Our new works generalize it to the embedding in two least significant bits, that is to say, combine two novel extensions of least significant bits embedding and the double-layered embedding developed in [16] with wet paper coding, respectively. The proposed schemes improve steganographic security and are less vulnerable to steganalytic attacks compared with original schemes with shared selection channels between the sender and the recipient.
An Efficient Identity-based Ring Signcryption Scheme
ID-based ring signcryption schemes (IDRSC) are usually derived from bilinear parings, a powerful but computationally expensive primitive. The number of paring computations of all existing ID-based ring signcryption schemes from bilinear pairings grows linearly with the group size, which makes the efficiency of ID-based schemes over traditional schemes questionable. In this paper, we present a new identity-based ring signcryption scheme, which only takes four pairing operations for any group size and the scheme is proven to be indistinguishable against adaptive chosen ciphertext ring attacks (IND-IDRSC-CCA2) and secure against an existential forgery for adaptive chosen messages attacks (EF-IDRSC-ACMA).
Multi-Recipient Signcryption for Secure Wireless Group Communication
Secure group communication is significant for wireless and mobile computing. Overheads can be reduced efficiently when a sender sends multiple messages to multiple recipients using multi-recipient signcryption schemes. In this paper, we proposed the formal definition and security model of multi-recipient signcryption, presented the definition of reproducible signcryption and proposed security theorems for randomness reusing based multi-recipient signcryption schemes. We found that a secure reproducible signcryption scheme can be used to construct an efficient multi-recipient signcryption scheme which has the same security level as the underlying base signcryption scheme. We constructed a multi-recipient scheme which is provable secure in random oracle model assuming that the GDH problem is hard, based on a new BLS-type signcryption scheme. Overheads of the new scheme are only (n+1)/2n times of traditional ways when a sender sends different messages to n distinct recipients. It is more efficient than other known schemes. It creates a possibility for the practice of the public key cryptosystem in ubiquitous computing.
Provable Security of Digital Signatures in the Tamper-Proof Device Model
We study security of digital signature schemes in tamper-proof device model. In this model each participant of the protocol has access to a tamper-proof device to encrypt hash values. This substitutes commonly used random oracle. In the tamper-proof device model we demonstrate security of a modified version of the GOST signature scheme.
Universally Composable Security Analysis of TLS---Secure Sessions with Handshake and Record Layer Protocols
We present a security analysis of the complete TLS protocol in the Universal Composable security framework. This analysis evaluates the composition of key exchange functionalities realized by the TLS handshake with the message transmission of the TLS record layer to emulate secure communication sessions and is based on the adaption of the secure channel model from Canetti and Krawczyk to the setting where peer identities are not necessarily known prior the protocol invocation and may remain undisclosed. Our analysis shows that TLS, including the Diffie-Hellman and key transport suites in the uni-directional and bi-directional models of authentication, securely emulates secure communication sessions.
Pairings on hyperelliptic curves with a real model
We analyse the efficiency of pairing computations on hyperelliptic curves given by a real model using a balanced divisor at infinity. Several optimisations are proposed and analysed. Genus two curves given by a real model arise when considering pairing friendly groups of order dividing . We compare the performance of pairings on such groups in both elliptic and hyperelliptic versions. We conclude that pairings can be efficiently computable in real models of hyperelliptic curves.
Last updated: 2008-08-28
Construction of Resilient Functions with Multiple Cryptographic Criteria
Uncategorized
Uncategorized
In this paper, we describe a method to construct (n, m, t) resilient functions which satisfy multiple cryptographic criteria including high nonlinearity, good resiliency, high algebraic
degree, and nonexistence of nonzero linear structure. Given a [u, m, t+1] linear code, we show that it is possible to construct (n, m, t) resilient functions with multiple good cryptographic criteria,
where 2m < u < n.
Cryptanalysis of a client-to-client password-authenticated key agreement protocol
Recently, Byun et al. proposed an efficient client-to-client password-authenticated key agreement protocol (EC2C-PAKA), which was provably secure in a formally defined security model. This letter shows that EC2C-PAKA protocol is vulnerable to password compromise impersonate attack and man-in-the-middle attack if the key between servers is compromised.
Cryptanalysis of Bohio et al.'s ID-Based Broadcast Signcryption (IBBSC) Scheme for Wireless Ad-hoc Networks
Uncategorized
Uncategorized
Broadcast signcryption enables the broadcaster to simultaneously encrypt and sign the content meant for a specific set of users in a single logical step. It provides a very efficient solution to the dual problem of achieving confidentiality and authentication during content distribution. Among other alternatives, ID-based schemes are arguably the best suited for its implementation in wireless ad-hoc networks because of the unique advantage that they provide - any unique, publicly available parameter of a user can be his public key, which eliminates the need for a complex public key infrastructure. In 2004, Bohio et al. [4] proposed an ID-based broadcast signcryption (IBBSC) scheme which achieves constant ciphertext size. They claim that their scheme provides both message authentication and confidentiality, but do not give formal proofs. In this paper, we demonstrate how a legitimate user of the scheme can forge a valid signcrypted ciphertext, as if generated by the broadcaster. Moreover, we show that their scheme is not IND-CCA secure. Following this, we propose a fix for Bohio et al.'s scheme, and formally prove its security under the strongest existing security models for broadcast signcryption (IND-CCA2 and EUF-CMA). While fixing the scheme, we also improve its efficiency by reducing the ciphertext size to two elements compared to three in [4].
The Random Oracle Model and the Ideal Cipher Model are Equivalent
The Random Oracle Model and the Ideal Cipher Model are two well known idealised models of computation for proving the security of cryptosystems. At Crypto 2005, Coron et al. showed that security in the random oracle model implies security in the ideal cipher model; namely they showed that a random oracle can be replaced by a block cipher-based construction, and the resulting scheme remains secure in the ideal cipher model. The other direction was left as an open problem, i.e. constructing an ideal cipher from a random oracle. In this paper we solve this open problem and show that the Feistel construction with 6 rounds is enough to obtain an ideal cipher; we also show that 5 rounds are insufficient by providing a simple attack. This contrasts with the classical Luby-Rackoff result that 4 rounds are necessary and sufficient to obtain a (strong) pseudo-random permutation from a pseudo-random function.