average case computational complexity theory - tspace

316

Upload: khangminh22

Post on 21-Feb-2023

0 views

Category:

Documents


0 download

TRANSCRIPT

Average Case Computational Complexity Theory

A the& stzbmitted in confoRnity with the teqpirements

for the Degree of Doctor of Philosophy

Graduate Department of Computer Science Univeffify of Toronto

The author has granted a ncm- exclusive licence allowing tfie NatidLi'braryofCauadato reproclace, loan, distn'bute or sen copies of this thesis in microf~m, paper or eiectronr-c fwnats.

The author retains ownership of the copyright in this tbk. Neither the thesisntwsPbstantialexctactshit may be printed or otherwise reproduced without the author's permission.

L'auteur a acco& me licence non excsive permettant a la Bibliotb&ue nationale du Canada de reproduire7 p&er7 distri'buer ou ~ ~ d e s c o p i e s d e c e # e t h ~ s o u s la f m e de microfiche/fikn, de reproduction sur papier ou sm format khtmnique.

L'auteur conserve la propri* du b i t &auteur qui prow cette t h k . Ni la thke ni des erztraits substantiels de celle-ci ne doivent &e imprim& ouanrlrementreprochritssansscw aatarisation.

Average Case Computational CompIexiQ Theory

Dodor of PhiIosophy, I997

Tomoyaki Yamakami

Graduate Departxnent of Computer Science

University of Toronto

Abstract

The hardest problems in the compl&Q class N P are d e d NP-complete. However, not d NP-complete

problems are equally hard to solve b m the average point of view. For example, the Hamiltoniau circuit

problem has been shown to be solvable det-y in polynomial time on the average, whereas the

bounded tiling problem stitl remaius hard to d v e even on the average We therefore need a thorough

analysis of the average behavior of algorithms.

In response to this need, L. Levin initiated in 1984 a theory of average-case NP-completeness. Levin's

theory deals with average-case NP-complete problems using polynomial-time many-one reductions. The

reduciiility is a method by which we can classif3. the distri'bntiond NP problems.

In this thesis, we develop a more general theory of average-case complexity to determine the relative

complex@ of all natural intractable problems. We investigate structure of redua'bilities, in-

cluding a bounded-error probabilistic truth-table reduaiility. We introduce a variety of relativizations of

frmdamental a m a p z e complexity Jasses of distribtttional decision problems. These I.elativizations are

essential when we attempt to expand oar notion of axerage polynomial-time computabihy to deveIop a

hierarchy above amage NP problems.

A- analyses are very sensitive to the choice of pmbabiIity distriiutions. We have o w that

if the input p robabw djstri'ibution decays expanentkdy with size, for instance, all NP-complete problems

are solved "fast" on the average. This phenmneftou does not reflect a signi6cant feature of averagecase

analysis. This thesis includes a thorough and* of stmctttraI properties offeasily computable distnintions

and kasiily sampIable distributions.

In addition, one may ask how we can extract the errsentia average behavior of algorithms independent

of the choice of pmbabiritg. dhtribotians, To arrsarer this question, this thesis introduces the new notion

of computabilitp, which expands the bouudary of feasii1e computabm (snch as

polynamial-time compntabiIity), and asserts the irrvariance of amagmwe cmpkity of algorithms regard-

less of which feasibIy computabIe distributions are chosen. This thesis exiunines the hardness of this red

compdabiE& and its StrncturaI properties

Preface

The theory of averageease NP-completeness came fora3Iy to my attention while I was a visiting scholar

at the Unive&& Ulm from April to August of 1991. In June of 1991, the annual meeting of complexity

theorists from the U n i d t Ulm and the Universitat Polithica de Catalunya was held in BarceIona Uwe

Schiining, the director of the Abteilrmg TheOretiSche Informatik of the UnivedZt Ulm, assigned to young

researchers the topics that would be exeensively studied at that year's meeting: average-case NP-complete

problems and local search problems. Six years before, L. Levin had presented his idea of averagecase

NP-completeness, and several important studies were done dong these lines.

I started reading fhse papers and technical reports and enjoyed discussing Levin's definition of "poly-

nomial on averagen with Rainer Schuler, who was hishiug his thesis on probabilistic computations. The

foundations of this thesis were estabIished during this time, and the results were presented at a conference

in New Delhi in December, 1992.

In June of 1994, I met Rainer SchuIer again at a coaference held m Amsterdam. He had with him a paper

which solved a prob1em we had left open m our 1992 paper. We soon started working together, refining his

key algorithm to construct hard sets which cannot be computab1e in fw31e time. These resalts were h e r

presented at a conference in Xi'an, C h h , m August of 1995 and are also induded in this thesis.

This thesis demands of little preparatory know1edge in the theory of computational compIe&y- Most

concepts are thoroughly defined in each section of this thesis or are self-expknatory-

I am extremely gratefuI to Stephen A. Cook for his hospitality and expert supervision. I thank him

also for his direction and support, without which I could not have come to Canada to pursue my PhD. degree. My thanks also go to my Send Rainer Schder who has been my collaborator since I visited the

Abteilung Theoretische Idormatik of the Unkmi& Ulm m 1991. I w d d like to thank Jie Wang and

Osamn Watanabe for helpfol comments and fmitfd criticism- SpeciaI thanks go to Ymi Gnrevich and ..41aa

Selman for his kindness and support. I am aIso indebted to Leonid Levin and Oded Goldreich fix helpful

commentsents I greatIy appreciate the input of my thesis committee members, Steve Cook, .b Borodin,

Alasdair Urqnhaa, Chariie Radro&, Ytrri G d c h , Anthony 1 Bonner, Rudolf Mathon, and Radford Ned

I thank my Eends Brian ESimn and Lnis Dissett at the University of Toronto for their kind advice and

encouragement. My spedal thauks also go to Eric HarIey and Debby F&pka for pointing out iqpcs and

grammatid errors in an eady mamscr@t.

My parents, Fbjb and Y o d b , hawe supported me emotionaIly and hanciaIly daring my studies m

Toronto. I also thank my grandmother, Nawo, fhm the bottom of my heart for spiritna guidance My great

appreciation &odd go to my h c i e Mitsne Normrra who ha helped me write this thesis-

Tamnto,Canada

May 7,1997

Contents

2 Foundations of Computational Complexity Theory 9

2.1 Introduction .............................................. 9

................................. 2 2 hdamental Notions and Notation 10

22.1 Logic ............................................. 10

2 2 2 SetsandNumbers ...................................... 11

22.3 Graphs ............................................ 13 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 22.4 F i and Infinite Strings 14

22.5 E'unctions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15 ..................................... 22.6 Asymptotic Notation 17 ..................................... 22.7 Probability Measure 17

2 3 Models of Computation ....................................... 18

2.3.1 DeterministicTuringMachines ............................... 19 ............................. 2.32 Nondekmbkkic Thing machines 20

................................... 23.3 Oracle W g Machine 22 ................................ 23.4 Alternating Thing Machines 23

............................ 23.5 Worst-CaseTime/SpaceComplexity 24 ....................................... 2 4 Randomized Algorithms 25

24.1 RandomInpatDomains ................................... 26

................................ 2.42 Probabilistic lSuing Micbines 27 ................................... 2.5 Worst-case ComplexiQ CIasses 29

.................................... 25.1 Computable hctions 29

2 5 2 Complexity Clanses ...................................... 30 ................................... 2.53 Worst-case Hierarchies 33

............................... 25.4 Polynomial-Tie Reducibilitie 35 ...................................... 25.5 CompIexityCores 36

2 6 One-Wayhctions- ......................................... rn 26-l Hashbctions ........................................ 35

..................................... 2.62 One-way hct ions 38

. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.7 ReleoantTheoris 42

. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.7.1 Feasl'bIe RedNumbers 42

. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.72 Kohogorov Complexi~ 44 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.73 Resomce-BoundedMm 45

3 General Theory of Average Case CompI&ty 47

3.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 47 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3 2 Distributio~~~axtd Density hct ions 49

. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3.3 A Notion of Easy-on-Average 53 ...................... . 33.1 Naive Dehitiou of Average Polynomial T i e 53

. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3.35 Levb'sDehition 54

. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3.33 Basic Roperties .. 58 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3.3.4 Different Characterization 62

33.5 Randomhctiom ...................................... 64 .......................... . . . . . . . . . . . . 3.4 A Notion of Domination : 68

..................... 34.1 Dominaton Relations and Eqgivalence Relations 68

................................... 3.12 hbd Roperties 72

.................................. 3.4.3 Randomized Domination 75

. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3.5 Distributional Decision Problems 77

.............................. 3.5.1 Average-CaseCompl~Classes 77

. ................ 3.55 Inclusionsandseparations ..... .-.. 83 .................................. 3.53 Another Characterization 90

............................................ 3.6 Furthdopics 97

4 Feasi'bIe Disttiiations 99

.............................................. 1 Introduction 99

...................................... 4.2 Computable DistnLbutions 101

.......................... 42.1 W o n of CornpntabIe Distrz'butions 101

................................. 422 RMShingsandRareSets 108 .............................. 4.2.3 FanIt-Tolerance 0fDistrriutions 114

. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.3 Normalization of Semi-Distributions 117

....................................... 4.4 SampIabte Distributions 120 ........................... 4.4.1 D a o n of SamplabIe Distributions I20

............................. 44.2 Inverbily SampIabIe Distn'bntiom 123

....................... 4.4.3 Ctosme Properties of Samphble Distfi'butions l27

.................................. 4.5 The P a m p = P-samp Qllestion 130

........................................ 4.6 UniversalDistribatiw 133

. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6.3.1 ReIativizedAvet(P. 3) 205

. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6.32 RelativizedAver(BPP, 7) 207 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 63.3 Relativized Aver(NP, F) 209

63.4 Relatrmzed * * . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Aver(PSPACE, 7) 214 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6.4 Average Polynomial-Time Hierarchy 217

. . . . . . . . . . . . . . . . . . . . . . . . . . . . 6.4.1 Average PoIynomial T i Hierarchy 217 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6.4.2 Sparse Interpolation Property 221

. . . . . . . . . . . . . . . . . . . . . . . . . 6.5 Average Polynomial-Tie Alternation Hierarchy 225 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6.6 Average Low Hierarchy 230

7 Quintessential Computability 233

7.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 233 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7.2 Rea Polynomial-Tie Hierarchy 235

. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7.2.1 TheNotionof"ReaICunderF' 236

.............................. 7.2.2 Real Polynomial-Tie Hierarchy 238 ............................... 7 2 3 Nearly-CE and Nearly-A: Sets 241

...................................... 7.2.4 CoIlapsing CIasses 242 ,

...................................... 7.3 hdamental Separations 245 ............................... 7.3.1 Construction of Hard instances 245

. . . . . . . . . . . . . . . . . . . . 7 3 2 Separation from "Quasi? Linear ExponentiaI Time 249 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 73.3 Separation from Advice Hierarchy 252

7.4 ImmtmItyandBi-I mnnmity . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 255 ............................ 7.41 Immune Sets and Complexity COWS 255

..................... 7.4% Bi-Immune Sets and Rggurce-Bounded M e i s m 257

7.5 ClostlrrProperties .......................................... 259 ............................... 7.5.1 Polynomial Time Reduu'bilities 260

7-52 PolynomiaIly h d e d W Operator ........................ 264 ........................... 7.6 Bounded Error Probabilistic PoIpomial Time 266

..................................... 7.7 RandomOracleSeparatio 11s 275

A Small Lemmas 283

References 289

List of Notation 297

Index 300

Chapter

Introduction

The new concept of the "automatic computing spstemn (a term coined by von Neumann) was proposed that

gave rise to cornputem in the mid 1940's. After five decade, compaters have come to permeate our society;

&- preslence spans the range horn wrist watches to weather forrcasting satellites orbiting the earth

The theory of wqmtdmd c m p k i t y has emerged as computer technology has advanced, and now we

face more diEculties than ever. When a problem is given, we must write a program or constract a drcait

to solve it. To minimhe the coat d solving the pmb- we mttst p d I y determine its complexity-

" C o m p I e can be measured in various ways, such as 'We running time spent by an algorithm," "the

memo y space useci for an alga- %he mnnber of basic operations made by an a f g o r i e "the number

of pt~~essors used for a circuit," and so on, Here we foeas on an algorithmic model of computation: tomst-

mse mmplexi~ t h m ~ deafs with the worst behaviors of aIgorithms, that is, the maxbd complexity of

algorithms when an adve~~ary chooses %adD instances. On the contrary, mmge-ease complexity theory

d p t e s algorithms by measorring their complexity on tire aoaage over all Instances.

lkaditid amapase analysis of problems has been performed to determitle the expected nmning

t imem~tapespaeeofalgoti thmstosolve~problernstmd~ ' cesinwhicheachmput

mstance occurs with a certain probabilitp. We have seat that many important problems, such as the traveling

desperscrn probIem aud the E h d t m h circuit problem, are categorized as the hardest to solve among ETP problems. The hardest problems in NP are d e d NP-compIete An NP-complete problems share the same

wimkase opmplex@, but they are not of the same average-case complexity. For example, relatively fast-

on-amqe d e t ~ ' ' "t - have been found for some famous NP-complete probtP_rrls. such as the

graph 3a- problem, and the Ead* circuit problem, under natnrany selected distriinticms.

..41thonghthenotionofexpectedrmmingtime/spaceissimpIe andhmitiq it hislimitations whenusedas

a base of a consistent d coherent theory and does not address the better understanding d the nature of mtractability d problem in both a themetical and practical sasa

In 1984, Ioeonid LevIn I601 presented a one page paper at the Spposimn on Theory of Computing,

STOC, p q a k g the novel idea of - an v,?npiBdtg measare. Levin demonstrated that

arandomEtedverSi~ofanNP-complleteptoblem,~rrmdomited~fitingp~iscOmpI~fota

randomized version of NP. This terse paper shed light on w h averagecase analysis should be. Early works

of Gurevich [36] and Ben-David, Chor, Gddreich, and Lnby [9] expanded Levin's original idea to establish

a coherent framework for averagecase compIexity theory. Siice then, numerous investigations have been

made-

This thesis tries to establish a general, consistent, and coherent theory of computational averagecase

complexity and to contriiute to its adva~~ement. In partl*cnlar, this thesis makes an important addition

to Levin's theory of average-case NPeomplekness by defining averagecase hierarchies founded on average

polynomial-time computable problems, anaIogous to the po1ynomiaI-time hierarchy. In this thesis, we study

the structure and properties of those newly de6ned hierarchies. We also emphasize the investigation of

distriiutions, which is a reced undertaking. The thesis carries out a thorough analysis of computable

distributions and samplable distriiutiom. The most innovative part of this thesis is the introduction in

Chapter 7 of tphtessential oomputability under a giveu set of distributions and its investigations. This

new concept enables us to discuss a wide range of subjects in a- compIexity theory. We use Kolmogorov complexity, resourcebounded measme, and random oracles to understand the true nature of

average behaviors of aIgorithms.

The thesis consists of eight chapters, each of which addresses a separate issue Specificallyt Chapter 2

provides the reader wi th the foundations of the theory of computational complexityt the fundamental notions

and notation, necessary to read the thesis. Most rend& come from the author's work (in collaboration with

R Schnler) on averagecase complexity theory [97,9Stll9], whiIe some new results are appended elsewhere

in the thesis. To avoid confusion, d t s (theorems, lemmas, e k ) with which the author was involved are

Iisted under Major C- at the beginning of each chapter with careful attniution. More detailed

explanations will be found bdm.

Easy on the Average. A naive idea of capturing the average behavior of a function f is given as for-

lows. For a distriitttion p, let as denote by ji the associated (pdabi&) density fWICCi4R The ftmc-

tion f is Uexpeeted polynomial on wverage" if there is a positive integer k snch that, for almost all n,

f ( z ) ~ ( x ) 5 nk, where p,, is the conditionaI distn'bution of p dehed on the sLrings of length n.

However, as dimmed m Section 33, this dehition has serious deficiencies, such as Mdnp the closure prop

erty under composit;on and lack& the property of machbAmdependence. It therefore cannot be the basis

for a consistent, fruitfd theory. In the a- setting, we view a decision problem as a pair consisting

of a set of instances and an input distribntian, called a d&dr&bd (decision) problmt or rrardomized

(&&ion) pproihu. The intended interpretation is that, for an aIgorithm which determines whether z E -4,

each instance z is given to the algorithm with a probabw spedfied by the distribution.

In contrast, Levin [a] d e d a function f po- on p-memge if there exists a positive real wmber 6

snch that the w o n &,,, [z[-Lf(s)dji(z) collvetges- This e q e c h t h is taken over the i&nite set

of a11 mnempw shings. Later h p a g b m [83j: pointed out that we can replace Levin's inhi& arpectation

with a series of- expectations, a an inpnt d e b n ) n ~ h . , - Ltrlsn f ( ~ ) ~ % ~ ( z ) being bounded by O(n), whete en is the con- d i s t n i o n of p on the of length at most TL In other words,

it is d a e r r t to check the -on over all strings of length at most n.

An intuitive characterization of Levin's notion of wynod on pamagen is given by Schap'ire [88] as

fbM t h e exists a polpomia psnch that, for every positive red number r7 F({z 1 f (x ) > p(r-lxl))) < l/r.

Here we remark that can be replaced by jr<,. - Based on Sehapire's formulatian, we are able to extend

M ' s pIyrmnial on p-mmye to t on p-aasoge for an arbitrary function t. Naturdly7 a distributional

decision pmbIem (A,p) is iden- as 4eu.sy8 un mcmge if the problem -4 is computed by a deterministic

'ILring machine which kaIts in po1yuomiaI time on paverage The dection of aU such easy-on-awfage

problems is considered an aveqpcax version of P by many d e r s and is denoted in this thesis by

Aver(P,*) (by AP, AvP, AverP, Aver-P, or Amage-P, elsewhere). This class is fundamental to Levin's

theory of a- N P a m p k n e s s - More generally, we can restrict ourselves to an arbitrary set 3

of distriiutions, and the notation Aver(P,T) denotes the collection of all easyan-average distrriutional

problems (A, p), w h p is taken h 3. Under some natural distn'butiom, several NP-compIete problems

are s o l d k 'Lfast on the a w n For example, the "satisfiabilitp problemn [a, the "graph 3-colorability

problem" [Ill, and the "Hamiltonian circait probIemn with edge probability 1 [13] are h d to be in

Aver@, t ) under some reasonable distriiolls.

On the other hand, an a v e q w a ~ counterpart of the class NP is the collection of all distn'bntionaI

problems which are paits of an NP set and a ferrsibip computabIe distxibution. This collection is denoted

m this thesis by Dist(NP,Pamp) (by W N P , RNP, or Random-NP elsewhere). Levin raised an inttigu-

mg qudm " Can all problems in Dist(NP,P+mmp) d y be "easy" on the average ?" Ben-David,

Chor, GoIdreieh, and Luby [91 gave the folIotrring answer: this is the case d e s s the nodehmb&ic lin-

ear -entiat-time class equaIs its deterministic comtapart. This thesis is motivated by Levin's open

Questioa Chapter 3 is devoted exclusively to mtrodudng Levin's theory of averagecase complexity and its

genemhtion.

To ded 6th the complexity issue, we generaIize the above tmo dasses and introduce the notion Dist (C, F) , which is the conection of alI pairs (A, p), where A E C and p E 3, and the other four fundamental notions

A = w , n A=(BPP77), A-~erlRp,T), and A=(PSPACE,T).

Input Disttr'butions. Here we wonld like to remind the reader that average-case analyses are sensitive

to the doice ofdistributiom, becatrse "aoerage poIycmiaLtime computabB'ity" is fonnded on the Mumior

d the distribntio~~ in gaestior~ The study of distnintions is thedore c r u d in averagPcase complexity

theory. In Chapter 4, we discuss the c o m p l e OffeasiIe distniutionsc In parti&, we shan focus on two

tgpes of distribntions: p o l y n d t i m e cornputabfe ~iutioons and poIynomiaEtime samplable distribu-

tions G d c h [36] cased a distri'bation p poEynomiaLhe conrputa6k if there exists a deterministic Thing

marhine M sach that k(z) - M(z,Oi)[ 5 2-' fm all nonnegative integers i, Een4lavid et al, introdaced p l y n o m i a ~ s ~ d i s t n ~ w h i c h a r e ~ b p r a n d ~ a i g o r i t h m s ( e a l l e d s a m p I i n g &

gnrithms [9] or gemmkxs [96j) wki& nm in time polsnomiaI in the Iength of "ontpbn By Pcomp and

P-satup, we denote the sets of pdymmiaI-time computable and samplable mpectidy- In Secti011~5,mshall~thatpotsP~timesamplable~ati0rt~arrprrciseIy~~asPPsetsto

compute detembWcaUy m polynomial time,

Another important notion in Levin% theory of average-case NPcompleteness is domination relaths

among distriiutions. When a distribution p majorizes another distniution u within a polynomial factor, we

SaY that Cr p0hJlWmidkJ dominates Y. More precisely, p polynomially dominates v if there exists a polynomial p such that p(lz1) - ji(z) 3 5(z) for all strings z. Polynomiadomination of polynomia-time samplable

distributions is closely related to the existence ofcryptographic oneway fnnctions. A (cryptographic uniform)

oneway function is a function which is easy to compute but hard to invert on most instances and is believed

to exist by many mearchers, &-David et ul. 191 first found this connection and showed that if such

oneway functions exist, then there is a polymmbI-time samplable distriiution which is not polynomially

dominated by any polynomial-time computable distriiution. In Section 4.7, we shall show that a much

weaker assumption, the existence of NP sets that are not neariy-RP, is enough to get the same conclusion.

Here, a set A is said to be nearly-RP if some randomized algorithm computes A on most instances, and it

behaves Iike a onesided bounded-error probabilistic madhe w most instances.

Moreover, if two distniutions polpnomially dominate each other, we say that both are polynornially

epiwdent. For example, every distriiution samplabIe dative to BPP sets in time polynomial in the size of

output is polynomialIy equivalent to some polynomiat-time samplable distriiution. Under the assmnption

P = NP, every polynomial-time computable distriiution is polynomially-eqtlivalent to some polynomial-

time samplable distriiution.

Average-Case ReduciiiEty. Chapter 5 focuses on a d e t y of averagxzw red~d~lities. For decades,

d e r s have made great efforts to achieve a better understanding of the structure and properties of

inEradd& problems. The term NP-complete wm coined to d e s a i i the most intraeta6Ie NP problems,

and mafly mterrsting NP-problems are declared to be NP-complete, that is, the hardest problems to solve

in polynomial time.

Levin's innovation is tbe invention d an a- version of such a completeness notion among

distriitttional decision problems. His notion of completeness: is based on worst-case polynomia-time many-

one reduciiility with an extra condition, the stxded d m i n d h condiEton for the reduction function,

which guarantees that the reduction maps more likely instants to more M y instances. Ee showed that

the "randomized bounded tiling problemn is complete for Dist(NP,P-comp) under this type of reductio~

Smce his proof of completeness, only a dozen distriitrdonaI problems have been fotmd to be complete for

Dist(NP, P-comp). Typical examples are: the Pandomized bounded baiting problem" [XI, the "randomized

bounded Post correspondence probIed [S], and the "randomized d problem for The systems" [I121

under p o l y n d t i m e many-one reductions We shall discus the issue of deterministic reductions in Section

5 5

In Section 52, we shall f o d y htmduce the (atrerage) poIynomial-time --One reductions and CUE-

t i d e their structural properties, Wang and Belanger [I121 de6ned polynomial-time isomorphism between

two distriiutiond decision problems and showed that all known complete problems for Dist(NP,P-amp)

are indeed pol&mi& ijOmOlpldc Section 5 3 will show that d tppl*caI d&dmtbd probIems are

CHAPTER I . INTRODUCTION 5

complete for Dist(NP, Pcomp) and also plynomially isomorphic to each other.

Incompleteness resnlts have been achieved by Gurevich [36j and by Wang and &hgr [112]. GnreYich

(361 6rst drew attedion to distriiutions of exponentially-small probability? so-cakd pot dis t r ions , and

demonstrated that no flat distribution makes a distriiutional problem complete for Dist(NP, P-comp) unless

NEXP collapses to EXP. We notice that the distriiution used for the randomized bounded tiling problem,

for example, is not dat. As W q and Belanger pointed out, if we restrict ourselves to on- pcdynomially

honest reductions, we can drop the assumption EXP # NEXP. We shall show that distriitxtions of

another type, d e d sparse distniutions, which were introduced by Gurevich [36], also do not make any

distriiutional problem complete for Dist(NP, P-comp) unless NP coIlapses to P. This incompleteness isme

will be discnssed in Section 5.4.

Another type of important reductiou is Yprobabilistic" or %ndomizedn reduction. In 1988, Venkate

san and Levin [106] used "random reductions" to demonstrate the intwtabiliw of the randomized graph

coIorabiIity problem- Later Ben-David, Chor, Goldreich, and Luby [9] introduced two more notions '(ran-

domized many-one reductions" and "raadomized 'bring ductiom." In Section 5.5, we shall introduce an

averageease version of bormdederror probabilistic truth-table reduciiility. Despite the incornphteness re-

sult of dat distributions, we are able to prove that the randomized bounded halting probIem with a natural

flat distriiution is also complete for DistcNp, P a m p ) mder these reductions.

Average-Case Hieratcbies. In worsase complexity theorg, MeyerStockmeyer's polynomial-time hier-

archy? {A:, !EL, II; I k > 01, has pIayed a central role in capturing the magnittlde of intractability of given

problems. Chapter 6 will discuss a hierarchical issue fiom the averagecase complscity point of view-

The d&duthd polynomial-he hiaorchy under T is an extension of the polynomial-time hierarchy

in which Af; and Ci are replaced with Dist(A1,F) and Dist(EL,F), respectively- We shaII show that each

C-level of the hierarchy under P-comp, Dist(C~,P-comp), has complete problems under poIyuomiaI-time

many-one teducfions.

The notion wiU be introduced of (pM-time ZMng) s s e l f r e d u c j i among distnintiond decision

problems. To determiae the membership z €?A, we recttRiveIy produce other instances y which are of

d e r size than z, and reduce the vestion z E?A to y €?A. Smce the size of instances becomes smaller7

these reductions tenainate in poIynomiaIly-many steps. In worstcase complexity theory, the stishbility

problem, SAT, is a mid exampIe of &-reducible problems- We shall show that most known distriiatid

problems complete for Dist(Zf,P-mp), k 2 I, are seIf-reducible- Whether an complete problems for

Dist(NP, P-eomp) are seIt-reducibIe, however, is an open question. A s an application of se&redua'bitityt we

shan show that DistlNp, P-aimp) G Aver(BPP, *) if and onIy if Dist(NP, P a m p ) G Aver(RP, *).

In Section 6.4, we shaII imsoduce another awmpcase anaIogue of the polynomial-time hierarchgr, called

themaage plpomid-time idermelry under a certain set of distn'bution, to c h d y distributional decision

problems which are hard for Dist(NP,P-camp). The hiemdy is b d t above Averlp,F) and Avet(NP,P) - - u -P- osing-

The model of aternating Tndng machines gives mother * - for the PotynOmiaCtime hi-

erarchy. I n s p i i by this characterization, we shalI introduce in Section 6.5 an menrge pipmid-time

rakmdm kemdy under a set 3 d distn'butions wing a model of alternatitlg T ' g m;urhinplc. Hterest-

ingly, each IeveI of the average polymmia-time alternating hierarchy is characterized by relativized Turing

computability relative to classes in the distai'butional polynomial-time hierarchy. As a d t , in contrast

to the worst-case sitnation, the alternating h i d y is unlikeIy to coincide with the average hierarchy m

general (of course, depending on the tmderIying set of distciiutions).

As an example, we shan locate the pmbabilirtic complexity class Aver(BPP, F) m the average poIynomiat-

time alternation hierarchy.

Quintessential ComputabiIity. In Chapter 7, we s h d shed Light on the collective behavior of distriin-

tional decision problems under a certain set of distri'butions, such as P-comp or P-samp. This a p p d is

new m average-case c o m p w theory and helps us investigate average-case complexity classes m terms of

worst-case complexity classes. More preciseIy, we shall focus on a class of sets S, called "reaI P under a set

7 of distnibati~ns,n which extracts the essentials of averageese complexity dass Aver(P, *) in the sense

that, for every distriiation p in F, the distri'bntional problem (S, p) belongs to Aver(P, *). In other words,

S is computable by some deterministic Tixring machine whose running time is polynomial on paverage.

We are particularIy interested in feasl'ble distributions, such as P a m p . Let as denote by P-,, the

dass 'T under Psomp." We retnrn to Levin's original question, Dist(NP, P-comp) C_?Arer(P, *). Now his question on simpIy be rephrased in terms of worst-case complexity classes as: "Is NP included in

P-,, ?" Based on the average poIynomiaI-time hierarchy, we further define real polynomial-time dasses,

{Ai3, EIF , l IL~ ( k > O), caIled the d p a M - - t i m e IdenmAy un&r 3. This hierarchy enabIes us to generalize Levin's question to any level of the real potynomia-time hierarchy under P-comp: "Is EL inchded

m hikmp ?"

We win show that, fOr evety integer k > 0, A'; E AiF C A; and CE E XiF for my set T of

distributions, where A; is the &-th level of the hear-srponentiaEthe hieratchg-; in particular, P C P3 E E

if k = I. Specifically, let us denote by Phmp the coIlection of sets compntable m poIynomial time on

average under every exponentid-time computabIe distn'bution, Using a notion of compIeJdt9. cores, we are

able to show that P-, cobpses to P. More g e n d y , we are able to prove that = A; and

C:!-Q)mp = CL for dl k > 0.

Section 5.6 wil l discuss hardness d t s ofthe average polynomiaI-time hierarchy under a set of polynomial-

time computable dktriibations, We have already seen the mdusions P G P-, 2 E. In 1995, SchnIer

[92] showed that both indnsions are truly proper nsing a diagonalization over polynomial-time compntabIe

"semi-distn'bttti011~-~ (Later he gave an dtanative proof based on KoImogo~~~ complexi@) We extend his

technique and show m 73J an men more p e COnseQnence: PkmP DTIME(P) fbr each fked constant c > 0. This remit will be extended to any level of the real polynomial-time hierarchy

under P-comp*

A similar technique again enables us to show that A:- has a hard set that is not in AS/m for each constant c > 0, whereAf;lf(n) mgeneralisthecaiIectionofallsets, eachofwhichcanbecompnted by

some A:-type machine with sorue &ice jh&n of length f (n). We note that the dass of sets computed by

non-uniform polynomial-size circuits is Btactly the mion of dl dasses P / nk, k > 0. It does not appear to be

simple to improve our d t to answer the open question of whether dl sets in Pkmp have polpomia-size

circuits. However, Schder [93] recently proved that if aII sets in P-, have polyeomial-size circuits, then

EXP collapses to the second l e d of the polynomial-time hierarchy. Hence, based on the common belief

that EXP is difFerent h r n the po1ynomiaI-time hierarchy, it seems unWrely that an sets in Pbmp have

polynomial-size circuits. These issues win be dhmssed in Section 7-32

Another m i d exampie of intsactabh sets, discussed in Section 7.4, is P-imwme and P-bi-immune

sets. P-immune sets are sets that do not contain any infinite? P-subsets in them, and P-bi-immune sets are

P-immune sets whose complements are also P-immune. We show that there are some non-sparse P-immune

sets m P k m p , but Pp-,-,, has no P-bi-immune sets. This fact exhibii the strunaral d i f f i c e between

Pkmp and the dass ET wbich hsr both P-immune and P-bi-immune sets. Udng the faa regarding

P-bi-immonitp, Pkmp is shown to be mat2 with resped to Ldz's resource-bounded meawe, where a

complexity class is o h Caned d l if it has pmeasnre 0. (Note that E has pmeasure 1.) As an immediate

conseqpence, if NP is included in Pbmp, then NP has pm- 0, and this consequence again contradicts

the popular belief that NP is not smaIL Along the same hes, Schder [W] showed that the --table dosure

of P h m p and the 'Ihring dosnre of P-,, have diflierent measures.

Section 7.5 wiII show that A~Pcomp is not dosed tmder polynomial-time many-one reductions, the ex-

istential opemtors, or the pmbabwc operatom. Hence, the dass P k m p , for example, is structurally

diflikrent fiom most of the weII-known complexity classes, such as P, NP, BPP, and PP. However, it is not

known whether PkmP is dosed under phonest xnany-one reductions. Notice that the cIdss A:-p, real

A: under P-samp, is dosed tmder ghonest polynomial-time many-one distriiutions. We shaU show that if

Phmp is not dosed under phonest polynomial-time many-one reductions, then there is a polgnod-time

sampIab1e distribution which is not po1ynomiaIIy dominated by any poIynomiaI-time computable distri'btt-

tion. Under phonest manyone reductions, we are able to show that there exists a pair of sets in Pmmp which are not redu~ib~e to each other, a so-cded hcmpmble pair.

The qrrintessential complexity class BPPF exhi'b'its another structure. Due to Ben-David, Chor, Gob

Mch, and Luby [9], the assmnption NP E BPPM~ implies the condnsion E BPPkmPT where t the dass of sets compntirbIe m polynomial time with nonadaptive queries to NP oracles. On the

other hand, Schnler and Watanabe [96l extended a resnlt of Venkatesan and Levin [q and showed that

the NP S?BPPMp mestion is equivaIeut to the NP a P P h m p - As shown by Ben-David d UL [9], NP PMp I d to the conclusion E = NE. On the other

han4 NP PmP yields the consequence P # NP. Hence, the N P question cannot be

easily solved in the non-relativized wodd. At this point, we have no prospect for anmmhg Levin's ~nestion

eitherafhmtkty ornegatively. Nowletnstrrmourinteresttoa~onofthisqnestiox~ 1~198'1,

BermdtandGill[8]introducedandi(~ofmndmn~to&mthat PipdifhntbNPin Ym& - .

r e M m d & & M a r e p r e d s e E y , i f a n o r a c l e s e t i s c h o s e n ~ n n d o m , t h e p r o ~ i t y t h ; r t P ~ ~

NPrektipetothiporsdeist I n M o n ? . ? , ~ W s 6 a I I s h o p p t h a f N P a n d P ~ ~ d y e x ~

(ie., NP P-, and PMp NP) in "most" relatlvlzed . * worlds. To be more precise, let us denote

p ~ ~ - c o m p a "naturaln reIativization of the class Pmp relative to oracle X. We wiIl show that, with g NPX, dative to random oracle X. pmbab@ity 1, Z ptx,, and Pp.-p

Chapter 2

Foundat ions of Computational

Complexity Theory

2.1 Introduction

The theory of compctatjonal complexity &st drew attention h m mathematicians as a weak notion of

recursive hctions. To measure the complsdty ofagiven problem, we use particular mdels of computatio~

such as lhring machines, cirdts, or PRAM'S to d v e the problem.

In this chapter, we shall define and explain most ofthe fundamental notions and notations in (worst-cuse)

computational cornplQiEy theory so that the nninitiated reader can read through this thesis without the help

of supplementary textbooks.

In Section !U, we shall cover the elementary notions of gnrpb, numbers, sets, and funcrions. The

basic terminology m pb&ify themy and bgk will be also i n t r odnd The thesis foIlows the standard

terminology often used m mathematics and themticat computer science-

We use - mcdrines = a model of compntatiou. h general, deterministic Thing machines compute

partial renssiae fnnctiom, but our interests lie only m resourcdxnmded computations, and we need the

notions of naming time and tape space of the Thing madhex The reader should pay carefuI attention

to the models we shall use m this thesis becanse ditkent modeIs Iead to different consequences. Several

variations of Ttxhg madines wiII be iatrodaced m Seetion 2.3, and many popuIar complexity classes, such

as P and NP, win be dehed in Section 25. The fiefd of rmrdomired a&iths has grown tremendously in the last decade and has found mimy

appIications because of their sfmplicitp and speed- We shan introduce the basic notions of nndmid

~ ~ p r o b o b i E i s r i c ~ m a d c i n e s , a n d m n d o m ~ i n S e a i a n 2 4 .

InSectionZ.6, trnitrd~/cnrcfimuwitlbeintroduced. HashfanctionsareausefuItoolindesi@ng

raRdomizedaIgo*

Section 2-71 will explain the themy dpd@md analysis initiated by KO and fietiman [55] m the early

1980's. The theory of Kolmogm compleaity also provides us with a succinct description of information.

We also cover the notion of rejowce-bounded K o ~ m 00mpIQity m m and Lutz7s resource-bounded

mensun? theory, which are popular in structaral c o m p I e theory.

For complete references, the reader may refer to [42,91,4,45,80].

Major Contributions. Although this chapter is introd-, a few results are included.

Lemma 2.5.7 offers a new characterization of all A - I d dasses in the poIynod-time hierarchy by

the model of semideterministic alternating Turing machks which w m polynomial time with constant

alternation.

Proposition 2.6.4 shows the existence of an NP set which is not nearly-RP, provided that strong oneway

h c t i o ~ 1 ~ exist-

2.2 Fundamental Not ions and Notation

We shall begin wi th terminology from mathematical logic and then explain mathematical notions and nota-

tions.- graphs, sets, numbers, strings, and functions. This section will include a preliminary introduction to

probability theory.

2.2.1 Logic

In propositional logic, we deal only with Bodean - a h which take d u e s 1 (truth) and 0 (fals-.

(Note that traditionally, in mathematical logic, 0 represents TaIsehood" and 1 represents 9mth.") The

terms are Boolean variables and the logical constants 0 and 1. As Iogioal connectives, we use the symbols

7 (negation), A (conjunction), and V (disjrmction). The set of (proposiEionol) f d o s is defined by the

f0IIowing claflses:

(i) every term is a form*

(ii) if a and @ are formnlas, then -(a), (a 81, and (a V B) are formalas; and

(iii) formulas are d&ed d y by clauses (i)-(fi].

Unless there may be confnsion, we freeIy omit parentheses from formnlasx eg, --a and Q A (B V -y)-

The negation of a Boolean variabIe w is sometimes denoted by iT for simplicity. A Boolean variabIe and its

negation are d e d literals-

Let a = a(z~,z2, ...,z,J beaformnlawithalIdistinctvariablesbeingecplicitlyexhr'b'rtedaszL,z2, ...,%.

We write Vm(a) for the set {z1,z2,. . . ,%). A bnth asignmd for a is a function u : Vm(a) + {T, F).

Given a truth assignment a, we de& an euaZMfion [a], of a on a m the foI1awing recmsive way

(i) mthecasethataisavadabIeq[~I~=Tifand~if~(o)=T;

CHAPTER 2. FOUNDATIONS OF C0MPlJTATIONA.L COMF'LEXlTY THEORY

(i) in the case that a is of the form (b A a), [a], = T if and ody if B], = T and B], = 2'; and

A propositional formula a is sati&.bZe if there exists a truth assignment a for a snch that [aj, = T. In this case, a is said to sdi& a. For example, the fomnla -(z V y) A (-2 V y) is satisfiable, witnessed by

the assignment a such that u(z) = o(y) = T and a(z) = F. A formula a is valid (or a tadology) if [a], = T

for any truth assignment u for a.

For a property Q, the notation VzQ(x) means that Q(x) holds for all elements z, and the notation

32Q(x) means that there ex& an eiemeut x dsfjtng QIx). The notation 3!zQ(z) means that there exists

the unique element z satisfying Q(z). For a property Q de6ned on an idhite set S, we say that 8(z) holds

for almost aU (or dmost may) z in S if the set (s E S [ Q(z) does not hold ) is finite In this case, we also say that 8 holds h o s t crrerywhere. The notation ;ZQ(Z) means that Q(z) holds for a h a t all z, and aa 3 &(a$ means that Q(s) holds for infinitely many s. Cleariy 7 and 7 are dual concepts-

Genedy, for a property Q, we write [Q] = I if Q is true, and [&] = 0 otherwise. For a set ST xs denotes

the chumdaistie fwrction for S that is d&ed as xs(z) = [z E Sj. (Note that "characteristic fnnctions"

here are different from those used in probab* theory.) For brevity, we also use the notation S(z) to mean

x&)-

2.2.2 Sets and Numbers

Sets. Intuitively, a set is a collection of objects, called its members or elements. The notation z E -4

expresses that x is an eIernent of A, and E is called the membership reIatioa The symbol 0 denotes the

enrpdy s d that contains no elements. We use the standard set notation {- I -1. For exampIe, the notation

{z 1 Q(z)) represents the set whose elements x s a w a property Q(z). For two sets A and B, we say A is a &d of B, symbolidly A E B, if every element of A is an dement of B.

For twosets Aand B, the intenedhof Aand Bis denoted A n B andis de6ned by A n B = {a I a € A A b~ B)- TheunionofAand BisdendedA~Bandisd&edbyAUB={aIa~A v ~ E B ) . The

set A-BdenotestheofAandBthatisdhdbpA-B={aIa€A A b 4 B ) .

The of A and B , denoted Ax B, is the set of aII ordered pairs (a, b) snch that a E A and b E B, where an

~ p a i r i s t h e s e t { a , { a , b ) ) . hcontrast,theset{~b)issometimesreferredtoasanunordered~ The

parer set of S is denoted by P(S) and is defmd as the cdIection of aII subsets of S, ie, P(S) = { A [ -4 E S).

For a set ST [Sll denotes the cmdinolity of S that intnitivey expresses the number of elemeats in S. If - 00. S is not finite, then let \IS\ -

Binary Relations. A 6inmy ddim on a set S is a subset of the btesian product S x S, LC, ((a, b) 1 a,Q E S). Conventidy, we write aRb when (qb) E R For a b i i relation R on S, rm say that R

is m j k h e i f & hoIdsfaraIIeIemeuts a € St andthat it is fnmifheif & & a n d bRcimply&Eorall

a,b,c E S. Moreover, a relation R on S is symmetric ifaRb implies bRa for all o,b E S; on the other hd,

Rb-if&aadb&irnplya=bfbraflqb~S.

CEAPTER 2 FOUNDATIONS OF COMPVWZT0NA.L COMPLEXlTY Z'EEORY 12

Nmnbers. Let Z be the set of aIl integers {- - , -2, -1,0, 1, 2, - -1, and let N denote the set of all nonnega-

tive integers, called RwnberS. The set of dl m f h d namkn {z I m, n E 2, n # 0) is simply denoted

by Q and @ denotes the set of all nonnegative! rational numbers. S i I y , the notation B denotes the set

of dl nol MnnbefS, and in mmlar, we denote by R? the set of an nornegative red nnmbers. (Remember

that the superscript + does not mean "positive.") We use the notation cx, to mean the inFnity, and let

P = @ u {oo) a d BO) = I U (00, -m). For the arithmetical operations + (addition) and nd (mnltipli-

cation), we follon the standard conventiox for any nmnbeR r E B and s lit+ - (O), r + oc = oo + r = m,

s - ~ = ~ . s = ~ , - S - O O = cm+(-s) =-00,andO-a,=oo-O=O. Moreover,weassmnethat -a < r

andr < 00 for anydnumberr E 16e

The absolute udtte of a real ntlmber is denoted [rl.

For any two real numbers a and b (a <_ b), Iet (a$) denote an open (dl intewd dehed by (a,&) = {Z E P ( a < z < b); let [a, b) and (a, b] be halfopen intefttab which are defined by [a, 6) = (a, b) u {a} and

(a, b] = (a, b) U {b ) , tespectivelp; and let [u, q be an closed intemui defined by [a, b] = (a, b) u {(a, b).

For a d number 2, let LzJ (pwr of 2) be the maximaI hdeger not exceeding z, and let rxl (ceiling of

2) be the minimd integer not d e r than x.

Lebesgue Measme. For a dosed m t e d I = [qb] of the line It, let III = b - a. Let S = {IC)&c be

a countable coIIection of closed intends on IIP. For a subset E of It, we say that S is a m e r i n g of E if

E rM I t . The Lebesgue onta mcnrmc of a set E, denoted mm(E), is d&ed by

If E is meaStrrabIe, its Lebesgue outer measme is caIIed its W g u e measure (or simply maswe) and is

denoted by m(E). Note that m([O, 1)) = 1. (Thus, m is a probability measure on the sample space [O, 11.)

It is weII known that, sPPDming the Oamn of there exists a noameaSnrabIe set (see, eg, [115]).

PolymminL and Logarithms. We are interested ody in pcdynomids and logarithms w i t h integer

ficients Fatapositiveintegerd,apotpmid (inn) of d@zedisafimctionp(n) ofthefonn:

CHAPTER 2. M)UNDATIONS OF COMPUTATIONAL C0MPI;EXITY THEORY 13

where each % E Z and w # 0- The constants a ~ , a t , . . . ,a are called the eoeffieicntr of the polynomial,

Epomdds are functions of the fotm 2p(*), where p is some poIynomial. In particular, we call a function a

linear--talifit is ofthe form 2Cf+dfm~econstants c,d E v. This thesis uses mainly logorithmJ to h e 2, and for the sake of convenience, we often omit the base and

simply write logz for log2 z. Whenever we ded with logarithms of rational numbers, we follow a special

convention: we define log z to be 0 whenever 0 < z < 1 to simplify the case-by-case description. For brevity?

we also write llog(n) for Llog2(n + 1)j and write ilog(n) for pog2 nl for all n E N. The notation log(k) n denotes k iterations of logarithms, namely, define log(') n = n, and log(k) n =

1og(logQ-') n) for k 2 1. Also Iet log' n = mh{k E N I log(k) n 5 1). The function log* n grows extremely

slowly. For example, Iog' 16 = 3 and Iog' 65536 = 4.

The kth ~ m m o n i c mmber, H ~ , is deiimi by 2, i. The binomial c o e m are defined as follows. for n, k 2 0, if n 2 k, then,

and if k > n, then (P) = 0.

2.2.3 Graphs

A dirrcted graph G is a pair (V,E), where V is a bite set and E is a b i &tion on V (i.6, a subset

of V x V). The set V is d e d a vcrtez set or node set, and its element is caIIed a node or oe&z The set

E is called an edge s d , and its element is d e d an edge. An undirected gmph G = (V,E) is a variation of

ditected graph whose edge set is a sgmmetrk &tion, For an undirected graph, we identi& two edge. (a, b)

and (b, a) and often write {a, b ) as an unordered pair.

We say that a node t is adjacent to a node s if (s, t) is an edge in a graph.

A ( f ini tc)@pafhflengt lrkhandestoanadetmagraphG= (V,E) isa(finite) sequence

(ao,q, ..., vk)ofnodsmNsnchthat s = m , t=vt,and((zi,e+l) ~EforaUiwi thO < i < k. Inthisease,

wesaythat thepathp ~ t h e n d e s v ~ , s , - . - , v t a n d a l s o t h e e d g e s (tl0,q),(g?~f2)~---,(~k-l,~k)- A n d e t i s ~ l e b a n o d e s i f k e x i s t s a p a t h p b s t o t -

ApathissimpfeifaIIndesmthepatharedisthct.

Given apathp = (m,q, ..., vt), a dpdhp' is a subsequence ofp; that is, for some i,j w-ith 0 < i < j k,p'=(vi,v++t,.-.,q).

We can natnrally extend the d e h k h of graphs and paths to in jh i te g m p h and in* patlrs. For

example, an infinite path h m a node s in a graph is an inhite sequence, rather than a bite one, starting

from s.

AgraphGr=(V',E)isasabgraphofG=(V,E)ifV' G V a n d F S E. G i i a s e t V r c V , t h e

snbgraphofG= (V,E) indrroedby V'is thegraphG'=(V',E'), whereE'= {(yo) E E [ y v E V'). Antmdirected~phiseonnededif~twonodesare~Iehmeachdher,

In a graph, a path p = (g,s, ..., uk) forms a qJe if (i) vo = uk and (ii) vo # ui for some i with

0 < i < k. A graph with no cycie is said to be acyclic

A fbmst is an acyclic, undirected graph, and a tree is a connected, acyclic, undirected graph. In particular,

the tree that contains no nodes is called the empty tree or nu22 kee.

A rooted ttee is a tree in which one of the nodes is distinguished from the others; this distinguished node

is called the root of the tree-

Let z and y be any nodes in a rooted tree T = (V,E) with root r. The node y is called an ancestor of

x if there exists a path from r to x which contains y. If y is an ancestor of z, then x is a descedmt of y.

(Note that z is an ancestor and descendant of z i e ) The node y is d e d a pwent of z if (y, z) is an

edge on the path h m r to s. If y is a parent of z, then z is a child of y. Any two nodes which have the

same parent are si6kgs- A node with no children is caIled a leaf (or e z t d node), while the o&a nomleaf

nodes are d e d internut nodes-

A m h t e tooted at z is the tree induced by the set of all descendants of x.

The degree of a node z in a rooted tree T is the number of children of x in T. The depth of a node x is

the length of the path kom the root of T to z. The height of T is the Iargest depth of any node in T.

2.2.4 Finite and Infinite Strings

An dphabet C is a nonempw, bite set. G i i an alphabet Z, a w d or string over C is a finite sequence of

symbols h m Z. The empty Jfring is the uniqtre string consisting of no symbols and is denoted by A. Let

us denote by !P the set of all sttings over X (of course, Z' contains A), and for the sake of convenience, set

9 to be C - {A), the set of dl nonempw strings.

In this thesis, however, we consider only the binary alphabet C = {O,1} (a string over {O, I} is oRen

called a binary Jtt.ingj because this restriction does not affect any of om. arguments.

The Iengtiz of a string z is the wmber of symbok in x and is denoted by lzl. For example, lOllOOl= 5,

and m partienlat, 111 = 0. For every n E N, let C (Zsn, C a , respectively) denote all strhgs of Iength n

(Iength 5 n, Iength 2 n, respectivelgry). We note that a subset of Z' is sometimes called a Impage over E. Fortwostringszand y, the -ofz and y is thestring~0nsistingofthespmbo~ofxfo11dby

the sgmboIs of y, and is denoted by zy (or sometimes x - y). For example, if z = 0110 and y = 1011, then zy = 01lOlIOlI. Given a shing s, SF denotes the set {sy I y E CL). For a string x and a natmd nrrmber

n, the notation zD is reemsivey defined by: z" = A, and flC1 = z - P for n E N

We assume the stadad orderon 2':

(Sort Iength-wise and then sort 1exicographicaIiy.) With respect to this order, z- denotes the pmkcessm

of z if one exists, and 2+ denotes the mazsor of z, Far exampIe, 0 1 W = 0111 and 0110' = 0101- This orderfng enabIes to identify strings wi th naturd nmnbers in the foU- fishiom let a = A, SI = 0, s2 =I, sf = 0 0 , ~ ~ n d s o f d hperal ,Iet s, be then-thstring(N.B. XistheOthstrfng) ofC in the

order. It is easy to see that 1%1= Ilog(n)-

CBUTE23 2, F O ~ A Z 7 O N S OF COMPUTATIONAL COMPLEKITY TBEORY 15

It is convenient to dehne inf;fiite strings as infinite sequences of symbols fiom C. We sometimes dl a

string m C a w e sfring to stress the finiteness of strings. For simplicity, Zoo denotes the set of all infinite

w- Wesaythat z i sapre f ; zo fy , symbo l idyz~ y, i f z s = yforsomestrings. Foras t r ingzanda

natnral number i with i 5 Izj, the notation z+j denotes the &st i bits of z, i.e., the string s such that

1st = i and s C z. For the sake of convenience, whenever i > lzl, set z+j = x. M e n n o r e , by z,i we

mean the string s such that z = zd-1s. Hence, z = zh-124.

Let f beafunctiononN Forase tSECm,Sis of density f(n) ifIISnZkll = f(k) f o r d EN

The cmpkmmt of a set A, SymboIicaIIy x, is Z' - A, and the symmetric dijference of two sets A and

B,symbolidy AAB,is(A-B)u(B-A). The~ointunionofAandB,symbolidyA$B,istheset

( O z I z € A ) u ( l z l x ~ B).

Any subset of (0)' or (1)' is called a tally set, and TAUY denotes the collection of all tally sets. A

set S is (poiynomidly) sparse if there exists a polynomial p such that IIS n CII 5 p(n) for all n E N By SPARSE, we denote the dec t ion of all spatse sets. By definition, TALLY C SPARSE.

Dyadic Rationat Nmnbers. A r d number r in the unit intervaI [O, 11 is uniqudy identified with its

Jhottest binary repmentation, ie., of the form

where all ~ ' s and bj7s are in (0,l) (the term "shortestn is necessary because, for example, the binary

repreSeatatiOnofthenmnberO2is0.1asdasO.0111---1--~). Weusethenotation(a,---aa.bt---bk.-+)2

to denote this ( M e or M d e ) b i i representation. This expression hdps as iden* a real number with

a pair of (bite or infinite) string ~ - . . U O and h---bk..- separated by ".", the delimiter symboL By

padding 0's if if, we can view r as an infinite string m COg.

Let us define dpdic rational numbers as rational numbers w i t h w e binary rrprr~ezltatiom. Here are

examples- 925 is a dyadic rational munber and is identified with the string 1001.01, but 23 is not a dyadic

ra t id number becatlse its binary representation is of the form (10.01001 - - -)2 and is hhite

In general, we wil l be using nary (partial) fimctions- For a function f, dom(f) (domain o f f ) denotes the

set of elements h m which f maps, and ran(f) (mnge off) denotes the set of dements to which f maps

Wesay that f is a(partial) fnnaionfromA to B (or f mapsfiomA to B), symbolidy f : A + B, i f

A = domlf) and ran0 B, and that f is a (partial) h c t i o n on A i f f maps from A to k A fnnrrim f is

me+m (or injedm) if, fm any two dements z, y E domV), f (z) = f (y) implies z = y, and f is d e d ortfo

(or srrjeetme) iE, for every dement y E ran(f), there exists an dement z such that f (z) = y- If a fimction

f is oneone and onto, then we caJi f a 6ij-ecPion (or &diue)).

Far a function f and an eIement y, in general, the notation f -'(y) (insretse image of y by f) denote the

CaAPTER 2. FOUNDATfONS OF COMPUTATIONAL COW- TEEORY 16

set {z E dom(f) I f (x) = 3); however, if this set is a singIeton (i.e, ll{z E dom(f) [ f (x ) = y)ll = I), then

by convention f -'(y) denotes the element z meh that f (z) = y. The lanrbd4 notation m X d d u s is a d tool for describiig functions by their dues. Based on each

d u e f(z) of a function f, the lambda notation "Ax. f (2)" denotes the function f itsetf. Here we shall see

some exampIes. The notation Xz.(clogz + d) expresses the fnnction f defined as f (2) = clog z + d for all 2,

and AdPhfd expresses the frmction f defined as f (r) = 2 e ~ ' + ~ for all t.

For two ftrnctiom f and g, provided that ran@) S dom(f), the composifh f o g expresses the function

h such that h(z) = f @(z)) for all x E dom@).

Wt : say that f mcymcyori;zes g, denoted by f 2 g, if dom@) C dom(f) and f (z) 2 g(z) for all z E dornk).

A function f is [weakfy) innmring (or monotone) if, for every pair of elements z, y E dom(f), z < y

implies f (z) 5 f (91, and a strict& innadng function f is obtained simply by replacing the above condition

f (x) 5 f (3) with f (2) < f (9). Srmikrly, we can de6ne (weakly) dareosing functions and s W y dccnasing

firnctions. A fimction f is un6arnded if, for every z, there exists au dement y > x such that f (y) > f(z). A function f on X' is caIIed tength-immsbg if [f (=)I > lzl for dl z E I?, and f is length-preserving if

If (z)l = 121 for an x E E'.

A hction f h m dom(f) to B is m v e z it, for any r, y E dom(f) and any real namba 7 E [O, 11,

and f is crmearre if we replace the symbol 5 by 2 in the above meqtrality.

A function f on X' is potynomiolty honest (phonest, for short) if there is a polynomial p such that

121 5 p([ f (2) 1) for dl z. ShiIarIy, a frmction f on F is e q m m t i d l g h w t (exphonest, for short) if there

is a constant c > 0 such that 1x1 5 2 ~ I f ( = ) l * for aII x.

aaditionally, a function f on Z* is d e d p o t ~ l y bounded @bounded, for short) if there exists a

polynomial p such that If (z)l p(b[) for dl strings z. A function f h m C to V is i s e d potjFomioay

aoonded @bounded, for short) if there exists a polynomial p such that f(z) p(W) for a?I x 1361. Note

that any composition of two pbotmdd fnnctions is also pboandd S i I y , t z p m d i d y bounded (exp

bounded, for short) hctions are defined by repking An) as above by an exponential Mn). A frmction f fiom dom(f) to B is pojitioe if f(z) > 0 fix all x E dom(n. Gnen a subset S of dome,

wesay that f i s p e e msif f(z) >OforaUs€S.

For any fimctions f and g mapping to F, we denote by f x g, f + g, min{f,g), and max{f,g) the

fimctiom deffned, respectively, a follows for all z, (f x g) (2) = f (z) - g(z), (f i g) (x) = f(z) + g(x),

*{fT91(4 = ~ { f ( & ? ( z ) ) , a d mar(f,d(4 = -{f(z),g(dI- For a function f h m N to IR?, f is neg@ibZe & for every po&tive polynornid p, it holds that f (n) <

for~ostaIlnatmalmtmbeft~

For~~aandb,thendationalbm~tbatthereerjstPan~ePatisfyiagb=e-a The

equidence relation of congruara moddo n, is defined as f 0 I . h ~ two integers a and b are congruact modtJo

n if n[(a - b), and this is denoted by a b (mod n)-

Let f beafmrctionhnN(orR) to$andtetr~B ~fbrewqreaLrmmberc>O,thereexistsanmnber

CILLPTER 2. FOUNDATIONS OF COMPUllATIONAL C O M P J m THEORY 17

E N (or z o E R) such that I f ( y ) - rl < 6 for atl y in N (or R) larger than than, we write lim,, f (2) = r.

Analogously, for an increasing function f from E' to W , the notation m, f (z) = r" means that

(i) f (2) 5 r for all z E C'; and

(ii) for every real nmnber s with s < r, there exists a string z such that s 5 f (z).

2.2.6 Asymptotic Notation

We aften use 0(.) (big oh), o(-) (little oh), a(*) (big omega), w (-) (little omega), and 8(-) (theta) as sets

of functions. Let f be a function from N to P. We f d y dehe five sets, O(f) , o(f ) , !2(f), ~ ( f ) , and

Wfl:

1. O ( f ) is the set of functions 6 such that, for some eonstaat c > 0, h(n) 5 c - f (n) for almost all n.

2. o(f) is the set offunctions h such that, for eoay constant c > 0, h(n) 5 c - f (n) for almost dl n.

3. Q ( f ) is the set of functions h such that, for some constant c > 0, c f (n) 5 h(n) for almost all n.

4. w(f ) is the set of functions h such that, for meq constant c > 0, c - f (n) 5 h(n) for almost all n.

To emphasize the variable n used for the fimction f, we aIso write O(f (n)) for O(f) and similarly for the

other four sets.

For example, An% E o(n2) but h!h2 $ o(n2); XnsZ/2 E w(n) but h . n 2 / 2 4 w(n2).

Definition 23.1 We dehe the foIlowing three notations:

TMitionall y, the notations O(f(n)), etc , are defined as *don-frmctiofls: the notation "g(n) =

O(f(n))," for example, means that g is m O(f(n)). In this thesis, we foI1ow this convention and l d y use the notations 0(-), etc , as if they are "fnnctionsOIISn As an example, when we write that n! = o(&(:)~),

we actualIy mean that the fhnction had belongs to O(&(:)~)-

22.7 Probability Masum

We begin with the hrmal dehitions of probability theory.

A sample space R is an underlying set- This thesis uses a subset of F' as a sample space R. A a-fieId (em amisk of a sampIe space R and a sabset F of P O satisfpfng the fbJJowing conditions:

(3WTER 2- FOUNDAXIONS OF COMPUTATIONAL COW- THEORY

(i) 0 E 1F;

(5) E E F implies E F, where Z = R - E; and

Any set in F is refkmd to as an event.

A pbcrbiIity ma~surr Pr is a hction fiom F to [0, 11 that satisfies the following conditions

(i) for dl set A E R, 0 5 =[A] < 1;

(ii) h[Q] = 1; and

For an went E, the notation Pr[a denotes the prvhbdity of E- A sapport of R is any Fset A for which

R[A] = 1.

The eondaioncl pn,buM@ of El given E2 is denoted by R[& I Ed and is given by

assuming that R [ E 4 > 0.

A p b & B y space is a triplet (Q, F, R), where (R, F) is a c-fidd and Pr is a probabiity measme defined

on the sample space Q. When R is dear fiom the context, fl may be omitted.

A coflection of events {&}iEIt where I is an index set, is independent if, for dl s u b S 1,

R[niEs&] = nips&[&]; or eqrddently, R[Ej I nips&] = R[E,.] for j E I- S i b , it^ is p h i s e indcpau*nt if, for any pair {i, j } C I, R[& n Ej] = R[&] - R[&,].

A (discmte) ntmbm urmdomk X is a fimction over the sample space 0 whose range D is either a W e or

countable h k i t e subset of B such that, for aJJ z E D, {w E R I X(w) 5 z} E E By identifping Xo with N,

we can introduce discrete random variables whose ranges are particular subsets of F. The eqected whrc or apcdatirm of a random variable X is denoted by and is defined by zzEQ 2 -

Pr[X = x].

In this thesis, we deal m=linly with discrete prohWty maacme on a &eId wi th a sample space R Cw,

and the notation Rk] wil l be be to denote the "tmifomn probabilItg. meamte For more details on

probability theory, the reader may refer to a text devoted to the subject, for exampIe [Ill.

2.3 Models of Computation

As a model of ucmmpntati~n,* we ffoco on l b h g moddncs which were introduced by A. 'Ihring and E. Post m the 1930's. This thesis uses the standard models of %ring machines with a bite namber of s--in-

tapes (ie, the tape has a leftmost square but is h&te to the right).

W e s6dl informdly me the term %Igo&hms" and %IgorithmicaIiy compnfab1e" in this thesin Although

thaeisnop~d~onf~theseterms,nestandonthe~rnm~~~knowna9Chmch'sTbesis,

t h a t a t g o d t h m p a r e d e ~ ~ ~ i ~ ~ ~

CHAPTER 2 FOUNDATIONS OF C O i W ~ 4 T I O N A L COMPLEXlTY TBEOBY

head

1 inpatape

loblolol I

1st w o k tape

2nd work tape

Figure 2.1: The hardware of a W g machine

2.3.1 Deta ' ' tic Tbrbg Machines

A k-tape off-line detaministic %+ng mnchine is formally a sextuple (Q, C, go, ACC, REJ, 6) which consists

of Q, a finite set of states; X, a tape alphabet with a special symbol for the blank; qo, an initial statq

ACC, a set of accepting states; REJ, a set of rejecting states; and 6, a transition function h m Q x Ck to

Zk-I x Q u ACC u REJ x {R, N, L)'. F v 21 mostrates the hardware of a lhrlng machine.

'k ing machhm are camdy caIled just mocirines. !I%e ftmction 6 is considered a progmm (or an

algoritlnn) ffor the madhe, and we often identify a Tbring machine with its program (or algorithm). A a n r ~ (or instwtrmeous desu+pfh) of a machine M is a description which contains the contents

of each tape, the position of each tape head, and the state of the machine, The M a l configwotimr is a

con@natkn m which the input tape contains an input, other tapes are blank, the internal state of the

machine is the initial state, and all head positions on tapes are the leftmost squares. An accepting (rejeding,

resp.) con- of M on z is a cdgtmt ion of M on z whose state is an accepting (rejecting, resp.)

state. A halting con- of M is either an accepting or a rejecting conSgnration, ie., a coxQumtion

h m which no other c o ~ o n s can be reached by the transition functio~

To d e ~ ~ l ' b e how the h e works, we need a concept of "cornputati~n.~

Dehition 2.3-1 (Cornpatation) A a m p t d h of M on input z is a (6nit.e or seqpence of

(i) it starts with the bit id confignration of M on z;

CIEAPTER 2 FOUNDATIONS OF COMPUWUONAL COMPLEXITY TEIEORY

(ii) each step h m a co&tmtion to another c o n f i ~ o n is made by the tramition ftmction; and

(iii) if 6nite, it ends in a Mting conSgmation of M on x.

An accepting (kjeding, resp.) c m p d d h is a computation which terminates in an accepting (rejecting,

resp.) configmation.

A deterministic ' k ing marhine M accepts an input z if there is an accepting computation of M on input

2; otherwise, M rejects z. Denote by L(lld) the set of aII strings which are accepted by M.

Let M(x) denote the output of a machine M on input z if it exists. The Nrming time of M on input z

is the length of the computation of M on x, and we denote by TmeM(x) the running time of M on z. In

the mse where the computation is not finite, we set Time&) = oo.

D-on 21.2 (Tiie/Space Comctii1e) A hct ion f on N is called time-GlltlSaUCtibk if there

exists a deterministic 'king machine M which, on input In, terminates a . d y f (n) steps are made.

A ftmction f on N is qmeconskudibiIity if there exists a detednistic W g machine which, on input In,

it marks the f (n)th square of the fitst work tape (among a bite number of work tapes).

2.3.2 NondeterminiPtic Turing machines

Another important model of computation is "nondeterministic" 'Ihring machines. A nondetffministic lbing

machine is a variant of deterministic Thing machines with the exception that the transition function 6 is a

map kom Q x 8' to P(x~-' x Q u K C u REJ x {KN, L)*). As for nand- ' ' tic ' h b g machines, we dter the dehition of "mmputation" to a set of "compu-

tations," a d e d "computation tree? A w m p f d h tree of .kt on input z is a tree whose nodes are

c o ~ t i o n s of M on x, i~ which the root of the tree is the initid configuration, and the children of each

node are such C O ~ O D S that are rea&abIe ftom the node in one step by the transition fnnction. Ail

codgurations foIIoning each configuration by a single application of the transition function is Caned nun- * . - detennartsttc Jloicts if the mmhr of such consgnrirtions is more than 1- .b accepting (rejectbg, resp.)

c t m p t d h is a path from the root to a leaf which ends wi th an accepting (rejecting, resp-) con@mtion.

The accepting aiteria of nondeterministic %rhg &es is simlt;tr to that of deterministic machines

and is determined by the existence of an accepting oompntatio~~, More precisely, the machine M aaq& x

if there exists an accepting computation of M on input z; otherwise, the machine z.

A nonde&mb&ic ' h b g machine for which the nrrmba of accepting paths on each input is at most

one is d e d uMncbigrmr [I03].

In general, mmapxse aunp1eJdtp measure is sensitne to the dehition of time-complarity of nonde

tmnbhtic 'Ihring machines (see [9Q, and we shodd pay carefd attention to the d M o n of the rmming

time of the machine when dealing w i t h nondeterministic compntati011~

In worst-case a m p l e theory, the nmning time o h nondeterministic 'Ihring machine is o h dehed

to be the minird Iength of allam@ngaxnptttaljonpaths Zone exiskg othempise, it is de6ned b he

ClUFTER 2 FOUM,ATIOPTS OF COMPlYZU7ONAL COMPLEXITY THEORY 21

I. Fix time/space constructibIe corn- bounds [such as "poIynd-timen or Yogarithmic-timev), we

may asmtme that all amputation paths on each input are of the same length, and by convention Iet the

rrmning time be the maximal length of any cornpatation path, since timeconstructible complexity b o d t guarantee that we can modifg any machine having this bound to a machine for which the length of any

computation path is d y t. This is explained as fd lows every time-bounded lhring &a is designed

in sn& a way that each has an internal dock (which does not access oracles), and this dock adjusts the

running time d the machine no matter what computation path it foLlows. We c d the mnr)l;nes esuipped

w i t h i r r t e r n a l c l o c l c s ~ ~ ~

Srnnmhg up, there are three m& of nondeterministic Turing machines together with their ruming-

time criteria-

(i) A model of nondetesministic Turing machines with traditional meaSzFement of running time, namely? the shortest accepting path if one exists, or eke L (or equhdently, taking the shortest rejecting path);

{iij A model of nondeterministic 'ihring machines with strict measurement of nmning time, that is, the shortest accepting path if one exists, or else, the longest rejecting path;

(ii) A model of ctocked nondekmhbtic 'Ihring machines.

As long as the runaing time of a mlchine on an accepting cornpatation path is bounded abcnie by s m e

thmm&m&& jhcrion (most timebtmded c o m p l e cIasses in worst-case cornpkxiQr theory satisfy

this condition), all these definitions are essentially equivaImt (negIecting constant slowdown). Smce average-

case c o m p l e theory does not reqnire this condition, the choice of a model is very important and often

leads as to &&rent mmeqnmces. Ia later chapters, we shaIt discuss the h o k e of models and possible

corlseqlleaces.

HistoricaIly, GciIdreich 1301 &st discnssed the average running t h e of nondeterministic ' k ing machines

and used a model of nm& ' ' tic 'Ittring mnrhines, the lengttts of whose computation paths are m e d

by some tim*bormded deterministic Thing e. His de6nition is actuany eqgivaleut to choosing model

(iii) as desaibed above Later Wang and Belanger [Ill], and also Schnler and Yamakami [97] prrsented mtetestfng resutts based on model (i). In paaicalar? W e r and Yamakami [97] constructed an a m q p

version of the (worsbase) polynomial-time hierarchy b a d on model (i), but the averagecase hierarchy

obtained here does not seem to be a pmper analogue d the worstcase hierarchy (it kds some properties Eke

@ =w). Int6isthesis,we&oosethemostgeneraImodel (i), eveuthoaghthemodelhnot seemto

p k d e the property that time-bounded nondeterministic compntations can be simulated by space-bounded

deterministic muhbe of the same m m p w .

D e f h k b 23.3 (ammine Time of Nondetermmtstrc * . . ThriqMachineP) Foranand' " * '

Ttrring marhine M, the rarcning Eime d M an inpnt x, Trmey(z), is dehed to be the length of the shortest

acceptingcomptrtationpathdM onxifoneexis& otherwise, it is- to be 1-

2.3.3 Oracle Turing Machines

To speed up the computation of an algorithm or to make it as accurate as posslIble, we need a supplementary

source of information which the algorithm can retrieve and ase. Such a source is called an om&, and a

'hing machine equipped with a system retrieving hfkmtbn h m an oracle is d e d an d e !bring

moehine An oracle Thing machine makes a p q to an o d e and receives its answer in a single step. We

start by descriiing those notions f o d y ,

An umck Zlaing mcelrine is a l b b g machine with the foIIowing additional devices: a distinguished tape,

aSLFcaIIedorocletopeor4uffy~andthreedistinguishedsbtes,QUERYtYES,andNO.Acomputation

(tree) of an oracle machine M with an orade (set) on input z is dehed in a way similar way to tbat for

"noa-oraJen 'Ihring machine except that it incorporates oracle qneries Initially the qyery tape is blank. If the machine M enters the QUERY state, then in a single step, M queries a string to the oracle which

appears on the query if this string bdongs to the o d e set, then M enters the YES state; otherwise,

M enters the NO state. h e d i a t d y after each orade query? the query tape becomes bIank,

We can easily extend the definition of made 'bring ma&iues with set macles (or o d e sets) into those

equipped with fundirm f. The d e machine has the QUERY state and the YES state; if it makes

a qgery z to an o d e , then the orade returns the d u e f (2) of the hct ion f in a single step and the

machine enters the YES state; at the same time, the head of the query tape is moved to the leftmost square

of the tape.

Smce orade lhring machines with the enrpfy ma& (ie-, the empty set) can be easily transIated into

non-orade '&hg machines (because we know the o d e answers), we often identify such oracIe machines

with non~rade oms. Zn this sense, without loss of w? we can view aoa-ode Tming machines as

a speciaI case of o d e Thing machines. Therefore, subsequent definitions wiU be stated only for oracle

machines without repeating s i m k dehitiom for nowrade machines

Definition 23.4 (Adaptive/Nonadaptive Query) An oracfe 'bring h e M is said to make

nonadoptirre queriej if, on each computation path, M produces a list ( d e d a qrrery Eist) of aIl strings which

are possibly queried befm the first query is made. Otherwise, M is said to make ndapfioe quaies.

A query list provides us with &dent i n f o d o n about which strings will possiily be queried in futtxe

cantpatations We remark that it is not necessary for an orade machine to query aIl the stzing m the query

Iist*

Let Acc(M, AJ) denote the set of (codes of) accepting compntation paths of M on mput z with d e

A, and Eimit;uly Rej(M, A, z) denotes that of computation paths. Let Q(M, A, z, y) be the set of

~qnededbpMarithoracIeAonmpatzanoompatationpathy. IfMisdderministic,thenwesimply

denote by Q(M,A,z) the set afaII strings queried by M aa inpat t wi th d e A.

By L(M,A) we denote the set of striugs accepted by M wi th orade -4, and we simpIy say that M with

oracIe A moprixs (or QCeCPfS) a sd B if B = L(M,A), For a machine M, MA(z) denotg the output of

a compnhtion of dB un mpat z- For a deterministic 'Zlg d M with an outpat tape (also called a

CZMTER 2, FOUNDATIONS OF COMPUTATIONAL COMPLEXlTY THEORY 23

tnrrrrsb),wesaythatM comprrteJafan&on f if f (x ) =MA(z) foraIIx€@.

B~ Tie$(x), we denote the rmLning time of marhine M with o d e A on input z. S i l y , Space$(z)

denotes the tape space used by M with oracle A on input x. Technically speaking, there are two possible

definitions for Space&(z) depending on whether the space of the query tape is counted. This possibly changes

the power of Wvized space-bounded compIexity classes, su& as PSPACE. In this thesis, we take all

tapes bdnding qaery tapes into consideration m order to measure the tape space used by the machine M

w i t h made A.

2.3.4 Alternating Turing Machines

The notion of olternnting mtdines was introduced by Chandra, Kozen, and Stockmeyer [21] as an

extension of nondeterministic 'Ihring machines.

Each m;lrhine is equipped with extra states, called V (aniversal) and 3 (existentiat). Each configuration in

a finite tree of computation for an a h m d n g 'Lhting machine is labeled as either inrittersul (V) or Qistentiol

0, according to the states of the machine. Next we define an accepting computation tree. F i we

recarsively determine the yes-amfigmutim

(i) a halting codgmation is a ye4conSgnration if it is an accepting configuration;

&) a non-halting 3-oonfigmation is yewdiguration if at least one of its children is so; and

(6) a non-halting V-configuration is a yescon&xation if all of its children are so.

For convenience, ccm6gmations which are not y-11s are called n o - m . m . An accepting

eomputafion bee Tt of M on input z is a subtree of a computation tree T of M on z satisfying the following

conditions

(i) all co&pratioxts m Tt are yea&pr&ions;

(ii) an existmW con@ration in Tt has one child node in T; and

(iii) a m i v e d configmation m T has all of its cbiIdren in T.

The machiue accepts an mput if there exists an accqtkg computation tree (equivaldy, the root of the

computation tree has a ges-eonsgmation); otherwise, che machine q,kds the mpta.

AIternation is the- over~attcompatationpathsfromtheroot toalea& ofthenumber oftimes

in which different Iabels of c o & p a t i o ~ ~ ( i i 3- or V ~ ~ i o n s ) change- Note that, by convention,

the initial con&tuatiion is asmmd to amtriiute the first aItematio11. For example, nondeterministic Thing

machines have 1- kame dl -ens are 3m@mations.

H e r e ~ ~ i m t r o d u c e a ~ ~ d ~ ~ m a r r h i n P I P . ~ e d " s e m i - d e t c ---*t?'

alternating Thing machines, which embodies deterministic computations reiatitre to some dternating ntriag

This notion is asefal fir descfi'b'mg the A-Ievel d the polyaomiaLtime hierarchy, for example (see

Section 25). A semi- - - -

a i t e r n a t i n g ~ ~ i s , m a g h I y s p e a k i n g , a a a t ~ ~ g

CE4FTER 2. FOUNDATIONS OF COMPUTATIONAL COMPLEXITY THEORY 24

machine which puts an additional mtriction on its computation tree. For simplicity, we assume that the

number of nondeterministic choices is 2 Then, any computation path can be encoded as a b i i string.

We IabeI each computation path with this code to chbgukh individual computation paths. For brevity,

write s? to denote the jth string of the set P in the standard order (NB. Om is the 0th string of P).

D M o n 2.3.5 (Semi-Deterministic Alternating lhring Machines) A semi-- . . . *-

ing machine is an alternating lSning machine with the following constraints for the 6rst three series of V-

and 3-con6gurations of the computation trees T, of the machine on each input z.

On input z, the machine starts with an 3state, and during this state the machine produces computation

subpaths. Let m be any such computation subpath and suppose that the marhine makes k queries By

the above coding scheme, the subpath fi is labeled by b = hb2 - -. bk, where bi E (0, I), I 5 i 5 k. Next

the machine enters an V-state in which it produces exactly 2* branches for some m 2 log k, each of which

is labeled by an m bit string c = C I C ~ where cj E (0, 11, I 5 i 5 m. Let be any such a branch

foIIowing m. Consider the current configuration, called cdicrJ, and denote it by 4, where I is the labeI bc

attached to the subpath plpz. At the next step, the machine &oases either an a t e or V-state, depending

on tJxe label 1 (= h bkcL k). The machine enters an 3-state if b, = 1 and c = s y for some j; otherwise,

it enters an V-state

AII critical configurations of the computation tree T= satisfy the fdowing two conditions on their h-

6e& with respect to yes- or n ~ ~ t i o n s : for any two critid codgaratiom dr and 4, where I = h - - - b k e , - - - k andl' = i4---b'&---t$,,, assnmingthat c = q a n d 8 = 9' for some j with 0 5 j < d { k , q,

A semideterministic Ttuing machine M is said to have k-ahm&ms if the maximal alternation of all

comptrtation trees of M is at most k -t 2 (because the &st two alternations are E d and &odd not be

counted).

We dehe the rrmning time of an alternating lhing machine as foIIows.

Dehition 23.6 (Rmming Time of Altematbg Zhring Machines) The rnnning time of an alter-

nating lhring machine on mpttt s is the mm;md height of accepting computation trees of M on % if M

acceptsqotherwise,thenmningtimeisde6nedtobeI.

2.3.5 Worst-case Time/Space Complexity

In womtxae complexity theory, the time (or space) annpiexity of an algorithm is often d d e r e d as a

function dehed an the natttraI numbers N because we are irrterested only in the instances of each Iength

which are hard to compute.

CZMTER 2. FOUNDATIONS OF COWUTATIONAL COMPLEXITY THEORY 25

Dchition 2.3.7 Let M be an o d e TLzring xuachine and A a set. Let t be a function on N, and let 7 be

a set of functions on N

1. The Tming macbe M with made A is d e d t-time bounded (or a t-time Twing mocldne, for short)

if Time&) 5 t([rl) kt all 2. S i l y , M with o d e A is t-space &nmded (or a t-spaw M g

mJdne) if ~pace$(z) I t(lz1) for all 2.

2. The oracle machine M is 7-time bouffded (or a 7-time 3kring machine) if M with o d e A A t-time

bounded for some t in 7. The notion of a 7-space %ing machine is defined anaIogonsly.

Here we use comt iond abbreviations if 7 is the set of polynomials, we say that bi is polpnnid-time

(or polgnomial-space) &taa&d, and simiIarly, if 7 is the set of exponentials (linear-exponentiaIs, Iogarithms,

resp.), M is called apimW-i.he (linew-ezponenticl-time, Iogarithmic-space, resp.) brnnuied

2.4 Randomized Algorithms

In this thesis, we would Iike to use two ditfkrent terms, mndomizcd M g maddnes and probabihtic l k n g

ma&nes, to cope wi th randomized atgorithms.

A nmdomizd Turing m& is a model of randomized computation and a variant of a nondeterministic

Turing machine with no accepting c&& Intnitively, we equip a machine with a s p e d rnechauh for

generating an u n b i i coin flip iu one step which determines the choice of the next consgarations. More

precisely, a randomized lbing mac?he is a 'Ihring macbiue with a distingnished state, d e d &-tossing

state, in which the finite control unit two poss'bIe next states. The (bite or inhite) -on

of a randomized 'Itring machine is determined by its input as well as by the outcomes of the coin tosses

performed by the machine-

We can view randomized Tmiug mdhes as pmtiol functions with two variables, one of which is a

*nsnaln inpd initially written on the input tape, and the other of which is a rnndom input (or m h seed),

that is a b i i sequence representing the oatoomes of the coin tosses. FoRotoing B k and Gurevich [12],

weshallformalizethigmodeIbelow*

A randomized Turing madtine M is equipped with an arndiarp semi-hdWe read-only tape, d e d a

random tape which may consist of an i n . sequence r of %domn bits ( i r E Coo). The head on

the random tape can move d y to the right and cannot stay at the same square after the machine reads a

symbol on the tape. The machine's a m s to the random tape corresponds to a coin-@, and we may say

that the machine pjps a (firir) coin when it accesses the random tape. For each infinite random sequence r,

let R e a d ~ ~ ( t ) be the initia segment of r that is rezrd by M on input z dming its computation. Note that

if a computation is bite, then Read~Jt) is aIso kite, but not c o d y . Let the sampk space ZIM(z) be

R e d that m is the -gne measure cm the Iine BL We d&e fhe probabiIity measme Pr on the sample

CHAP17ER 2 FOUNDATIONS OF COWUTA?IONAL COW- TREORY

space QM ( x ) as foIloftrs: for any event E &(z),

by identiEging [0,1] with !Em. Throughout this thesis, we use Pr[-1 to denote this probabii m m In

if event E consists only of bite strings, then

Abusing notation, we let RM denote the set ((2, s) I s E RM(x)). (Note that Ras is not a sample space.)

We shall dehe the nmning time of the machine M on input z as foIlows. For any input (z, r) E Z' x Zoo,

Iet TM(x, r ) denote the time taken by M on mput z and random input r. Whenever M does not halts, let

TM(z,r) = ao. Next Iet Tlme~(q -) to be the (partkd) function &om Rna(t) to N such that T&r) = T i e M (z; &adM&)) for d r E p. For convenience, if (2, r') RM, then mme~(z ; f ) is undefined.

Unless stated 0th- the notation ks.TimeM(z; s) is used to mean the total function defined as above

h r n I'na to N&.

For each input s to a randomized ntring madine, let US denote by M ( z ) a random variable over the

sample space RM(z). Let Q be a property on {O, 1). Then we denote by Rr[Q(M(z))J the pmbab'ity

that Q(M(z ) ) h01& For the sake of coLwenince, we also use the notation M ( q r ) to mean the output

of a computation by M on mput z with random seed r when r is m OM(=). In the case where r $?

flM (x), let M (q r) be undehed. Thus M(z; r) is a partial fimction The notation Pr,[Q (M(G s))] and

Pr.[Q(M(z;s)) I s E QM(=)] are to denote Raa[Q(M(4)]-

2-4-1 Random-Input Domains

B k and Gnrevich [U] created a general h e w o r k for the average analysis of randomized algorithms based on the notion of ''dil;iti~13~.'' Here we take a simpMed approach.

D-on 2.4.1 ( B a n d o ~ ~ t t t Domain) A subset r of X' x Z' is caIled a mzdbm-input &main

& for z, s,s' E !Em, (z, s) E r, s' 5 s, and s # s' impIy (q 4) 4 r. For each string 2, we set

r ( ~ ) = (S 1 (z,s) E r).

It is important to note that r(z) consists onIy of finite strings and that it may not be a sample space

with respect to Pr became Rp(x)] 5 1.

Dehition 2-42 (Raritp Fbctions) [I21 Let I' be a raudom-input domain. The m+tg /undion of I' is denoted by Ur and d&ed by

1

CBWlER 2. FOUNDATIONS OF COMPUZUTONAI, COMPLEXITY THEORY

In other aords, l /Ur(~) = Prr(z)]. For a raadomid 'Ihring m;vrhine M, let rM = {(x, s) € EM I s is finite}. In particular? when r~ is

almost total, there is no need to Setentiate bemeen r ~ ( x ) and the sample space R&) since R[r&)] =

R[RM(x)] = 1.

De5ition 2.4.3 (Random Mctions) A rmulmn j imdm f is a function from a random-input domain

to P . A random function f wi th its random-input domain I' is called crlmost to td if I' is almost total and

R.V(z,s) < oo I s E Rnr(x)] = 1, where the subscript s in this equation emphasizes the random variable

over RM(z). For a random function f with its random-input domain r, the (dd) ezpatotimr of f

on input Z, symbolically E.[~(x,s) [ s E r(z)], is dehed by

For example? the partial function )cu.TimeM(qs) for a randomized Turing machine M is a random

function fiom RM to N.

2.4.2 Probabilistic Turing Machines

The notion of probabilistic 'I\aing machines was proposed by de Leeaw, Moore, Shannon, a d Shaph [a] in 1955. Pioneer works on probabilistic 'Ihring machines were done by Gill [%I and Santos [86j.

A pmhbilistic Zhfng m c h e Y is a randomized 'lhring machine with the foIIowing accepting criteria:

for every z, either R. [M(q s) = 1 I s E R&)] > f or R,[M(z; s) = 0 I s E fh&)] > [28]. We sap that

M omptrzifPr.[M(z;s) = 1 I s E QM(x)] > $and M mjedszifR,[M(z;s) = O I s € QM(x)] > $. By this definition, the raritg fnnction tlr, of rar is bounded above by 2 because

For a prolmbilistc 'Ihring machine M, let L(M) denote the set of dl strings which are accepted by M. Wesay tha tMnmgnkesDi fD=L(M) .

Let D be the set recognizable by a probabilistic Thing machine M. The arm pmbob*ly of M for D is

the hc t ion e~ defined by

esr(4 = Pry[M(4 # m(4,

C&APTER 2. FOUNDATIONS OF COMPtl'ZUlONAL COMPLEXITY THEORY 28

Another type of meamre was given by Gill [q. For a probabilistic 'Ihring machine M, let D be the set

accepted by M. Set

min{n I Rs [M (z; s) = xD(x) within n steps I s E rrr (x)] > f ) if one exists, T'i'iea(z) =

othcmirk.

Roof. Let t be the error probabilitp bound of M . Hence? 0 5 I < Let c = 1 t -2c and c' = c'I6. For

simplicity, write h(z) = E.rieM(r;s)* 1 s E ru(z)].

Let x be an arbitrary string- By Lemma A.1,

This is eqttivalent to 1

R,[TimeM(r;s) > c' h ( ~ ) ~ * I s E I)M(z)~ < ;.

Let D be the set accepted by M.

The probabiIity that M(z;s) = D(z) in time d - h(z)ll% ip by

R*[M(z;s) = D(z) m time c' - h(z)'/' 1 s E ru(z)]

By the dehition of T i ( z ) , we condude that Tiieb(x) d - h(z)'I6, that k

CEAPEZ 2. FOUNDATfONS OF C O W ~ O N A L COMPLEXITY THEORY 29

De&rition 2.4.5 (Time Cornpkity) A probab i ic Thing mslrhine M is Caned t-time 6ounded (or

a t-time %ring machine) if Tme&(z) 5 t(lz1) for all x. For a set 7 of functions, M is 7-time bounded (or

a 7-time machine) if M is t-time bounded for some t € 7.

2.5 Worst-case Complexi@ Classes

For years, theoretical computer scientists have been interested in resource-bounded computations and have

studied their complexity and structuraI properties. In this section, we shall review central concepts in the

theory of worst-case complexity.

2.5.1 Computable Fnnctions

One of the most natural concepts in worst-case complexity theory is "polynomial-time comptttabilIitytyn .A

function f on C is polynomial-time mputuble (P-computable, for short) if there is a detc ' - tic Wing

machine with one input tape and one output tape (ie., a transducer), which computes f in time polpornid

in terms of length of the input. Denote by FP the collection of all polynomial-time computable functions.

To study algorithms on difFenat objects, such as graphs, sequences, circuits, ete, we use an encoding of

objects into stringq the encoding must be effective and same. In this thesis, we use Regan's paring function

[a as the basis of encoding and extend the function to mnIti-functions. Formdyt a patiing jhdion is a

bijection h m C' x C' onto C'.

F i d that the stat ion z- demtes tho predecessor of string z in the standard order on C' nnless

z is the empty string.

D M o n 2.5.1 ( P e g bction) The fimction (-, -) from C* x C' to Z' is defined as f o U m for

an pairs (x, y) E C' x E',

where d(X) = A, d(0z) = OOd(x), d ( h ) = lld(z), i2[Ox] = Oh, and i&] = 102 for all z.

Below we list without proofs severaI important properties of this paring frmctiow

(i) (,) is monotone, Le, x I d and y 5 y' imp17 (z, y) < (z',y').

(ii) (, ) is compntable in linear-time m the lengths of z and y.

(v) For any maeasing fimction f on C, Cf(x), x)- 2 Cf(z-), z-) fix all z.

CHAPTER 2 FOUNDATIONS OF COh4PUlZXTONAL COMPLEXITY THEORY 30

This paring frmction is recursively generalized to a bijection fiom (Em)' onto C* as (zr ,zz, -. - , zc) = (z~ , (z~ ," ' , z~) ) . It is clear-

2.5.2 Complexity Classes

A annpkitg dass is a cokction of subsets of Z*. For a complexity dass C, the complement of C, denoted

by~7isthecollecti~udsetsSsoehthatC'-SisinC.

Dehition 2.5.2 For a function t on N, let DTIME(t), NTIME(t), and DSPACE(t) denote the c h s

of all sets recognizable by deterministic t-time, nondeterministic t-time, and deterministic t-space Turing

machin=, respectiveIy. Similarly, Id BPTIME(t) denote the class of sets recognizable by t-time bounded

probabilistic M g machines with bounded error probabii , and let -(t) denote the couection of

sets computable by one-sided error, probabilistic (LC, random) t-time lhring machines. For a set 7 of

frmctions, let DTIME(7) = UtETDTIME(t). In a similar fashion, we define NTIME(T), BPTIME(T), and

DSPACE(7).

Using these notations , we can introduce several important complexity classes. For deterministic classes,

we use the following basic comphiQ classes:

2. E = ~ = ( 2 ~ ( " ) ) (linearexponential-time).

A set S in P is said to bepoEynmniodEime cmnputa6le (P-computable, for short). Srmilarly, we use the ter-

minology czponnztial-time mmput4ble (EXP-comp~le , for short) and linem-eqonentiol-time amapuhble

(Ecomputable, for short), respectivety, for sets in EXP and m E. For spacebounded c o m p I e chsses, we trse:

I. PSPACE = DSPACE(~*(')) (polynomial-space).

2 ESPACE = D S P A C E ( ~ ~ ~ ) ) (hear4?xponentiaspaCe).

It is woah noting that PSPACE is dosed iarder oomplement, ie, PSPACE = mPSPACE. For nondeterministic dasces, we set:

CaAPTER 2. FOUNDATIONS OF COMPUZXZTONAL COMPLEXlTY THEORY 31

-y, NP-sets have the foUowing -on by logid terms: a set A is NP if and only if there

exists a set B E P and a p l y n o d p such that A = {z I 3&l= p(]z i ) (x? y) E B]}. For probab ' ic cbsses, let:

1. RP = R3IME(n0(')) (random po1ynomial-time).

2 aE = RTIME(no(')) (random hear-exponential-time).

3. B P P = BPTIME(~O(~)) @ounded-ef~)r probabilistic polynomial-time).

4. BPE = B P T I M E ( ~ ~ ( ~ ) ) (bounded-r probabilistic hear-exponential-time).

The class ZPP (zememr probabilistic polpod-time) is defined by RP n co-RP.

It is important to note that DTIME(O(n)) # NTIME(O(n)) [82]. This is the only separation r e d t we

have known until now.

Other important comp1exitp cIasses are U P and PP. The chss UP (unamhiious polynomial time)

is defined by po1ynomiaI-time unambiions Thing mnrhines. The probabilistic class PP (probabiIistic

po1ynomial time) consists of sets which are dehed by polynomial-time probabilistic Thing machines whose

error pmbab'ity is < 1/2. ActuaIly? the value 1/2 can be replaced by any real number c sa-g 0 < c c 1. By dehition, P E UP E NP and P RP BPP E PP. Using the notion of orade lhriag machines, we can introduce ~ ~ e d e ~ ~ p l e z i t y b. For a set

S and a c o m p l e ekss C, the notation CS (C relative to S) denotes the dass naturdy obtained from

the dehiti01t of C wi th the help of S as oracle, and CD (C relative m V) denotes the union of all cS for

an. S E D. AU c o x n p l e classes defined above ate m d y dtivized wing the notion of oracle %ring

machines, such as pA, w, B P ~ , e, UP, ete

As mentioned before, the empty oracle set does not change the computational power of the %ring

m;lrhines; thus, we have = P, NP@ = NP, e tc We remember an important r e d t due to Zachos (see,

eg.? [lq) that BPF = BPP. Let #P (pronounced "sharp P" or "number P") be the coktion of all functions on !Z' which are

computed by polynomial-time countkg Thing machines. It is not Wdt to see that f is in #P if and

ody if them exist a set A E P and a polynomial p such that f(x) = Il{y E WM) I (z, y) E A}II for dl z.

The counting of soItrti011~ is relevant to probabilistic computation- The following redt shows a relation-

ship between these two notions.

Root Fmt we show the inclusion PP E F. This inclusion is d y seen as fo11ows. For a set A E PP,

there is a pmbddistic %ring machine M- We modify M so that all amputation paths of M on each input

z are of the same length, say p(z). The success probabw of M on z is eqasl to the ratio of [[Aa(M, z)1[ to Hz). Let us d&e f (z) = l~Acc(M,z)[ hr dl z. Obviously f E #P. U 5 g this f, we can determine in

p d y n d time whether 2 - f(z) > !2P(=), which means z E A. A beI- to Pf E p.

CaAPTEIl2. FOUNDATIONS OF COMPUTATIONAL COMPLEXlTY THEORY 32

Next we shall show the other mdnsion that #P E l@'; thus, we have P # ~ C pPP.

Take a #P-function f. There exists a polynomial-time nondeterministic Turing marrhine M such that,

for every z, f (z) equals the nnmber of accepting computation paths of M on inpa z. Take a polynomial p

and assume that the number of non- amputation paths of M on input z is exactly 2 ~ ( b l ) (this

is always true if we pad extra nondeterministic, rejecting paths).CIearly 0 5 f (z) 2p(lZl) for aII z.

We then define the set A = {(q,z) ( IIAcc(M,z)ll 2 i) so that:

f (z) = max{i E N I (si,z) E A).

A b i i search technique hdps as compute the d u e f(z) detc 'SticalIy by simply querying polynomidly-

many strings of the form (si, 3) , 0 5 i _< 2p@4), to or& -4. Therefore, we have f E F'PA.

In the rest of the proof, we mast show that A is in PP. Let us d e h e the following randomid 7 h b g

machine N:

begin randomized algorithm for N

inpat ( ~ i , 2)

if i = 0 then accept and halt

generate a bit i (b E (0, I)) at random

i f b = O thensimulate M onmputzandhalt

generate a string y of length p(IzJ) at random

(assnme that y is the jth string in D'(lzl), where 0p(IzI) is the 0th string)

if j c 2p(Irl) - i then accept else reject

end.

This machine N obviously nms in polynomial time because M does so. It is also easy to see that, for each

(sit 4, 1

Hence, A E PP. 0

A set S is m @P (pronounced "parity Pn) if there exist a poIynomiaI p and a set A E P such that, for

each z, z E S if and onIy if [I{y E CP(lrI) I ( q y ) E A}([ is odd [81]. We remark that UP s $P. A set

S is near-testable if there exists an f E FP such that, for all nonemptp strings z, f (z) = xs(z) + xs(z')

(mod 2). Let NT denote the coIleaion of all near-testable sets [31]. It is known that P C NT G $P and

NT is also indnded in En PSPACE [31]- It is also known that $Pp = $P [IS] and NP s eP [104.

The class P/poIy consFns of all sets A such that there exist a pbounded function f h m N to C' and

a set B E P satk@ng A = {z [ (2, f (kt)} E B). This class is a h known as the colIection of all sets

computable by (non-nnifonn) fhdies cdpolsrnomi&ize cimits.

A s e t S i s m A P T ( o b n o s L ~ ~ } [ ~ ] i f t h e r e ~ a p d p n o m i d p a n d a d e t e r m i n i s t c ~

machine M which accepts S snch that the set {x 1 Tme& 2 p([z[)) is sparse

CHAPTER 2. FOUNDATIONS OF COMPUTATIONNU COMPLEXITY THEORY

2.5.3 Worst-case Hierarchies

We shall define several important hierarchies m 7Aorst-case compIexity theory.

DeSnition 2.5.4 (Polynomial T i e Hierarchy) [72, 1Ol,ll8] The pol^-time hi&y con-

sists of &e following compl@ classes: a = Et = = P; Af; = p:-'; Xf; = e:-,; and

= co-zf; for k > 0. We also use the cumdative polynomial-time hierarchy PH = Ut2,, A:.

Note that Ef;uTti c A&, XpH, n c l for all k > 0 (see [IOlj). It is known that if NP BPP, then

NP = RP [52j and also PH = BPP [123]. A recent achievement is Toda's theorem [I021 that PH G pPP.

For a function f on N and a set A, let ~ ~ [ ~ ( f ( ~ ) ) 1 be the coltection of sets B which are computed by a

polynomial-time deterministic o d e Thing machine M with oracle A such that the number of queries by

M on input z is bounded above by c - f (n) + d, where c and d are constants depending only on M. For a

dass C of sets, set pC[O(f(R))I to be the union of all P~[O(~("))I for every A E C. In particular, we write 8; for pC5-l[oclo~)l-

Moreover, we define two hierarchies over E and EXP as foIIows.

Dehition 2.5.5 (Exp~nentiaE~me Hierarchies) cf. [4j

We next consider alternating Thing machhs.

Deanition 2.5.6 Let ATIME(t(n)) be the dass of ail sets which are computed by alternating M g

machines in time t(n). We also d e e two alternation-bounded classes. Let ATIME*(~(~), t(n)) be the

dass of alI sets comptited by semi-deterministic alternating Turing machines with at most k(n)-alternations

in time t(n). S i i I y , let ATIME'(~(~), t(n)) be the cIass of a11 sets computed by alternating

mdbes, starting with existential states? with at mast k(n) alternations and in time t(n). For sets K

and 7 of fimctions, we defme ATIME'(K, 7) = uLElh. utfl ATB@(k(n), t(n)), and ATIME*(F, 7) and

ATIME(7) can be d&ed simiIar1y.

All three hiexarchies introduced above are charaaerized by alternating Thing machines with constant-

alternation.

C&APTER 2. FOUNDATIONS OF C O M P M O N A L COMPLEXITY THEORY

Proof. Here we show ody the claim for A:. For a set A E A:, take a deterministic oracle Turing m;trhine

M which is ptime bomded and an oracle set B E Z:-, (ii k = 1, then let B = 0) such that A = L(M, B),

where p is an increasing polynomia. Smce B E C:-,, there are k-aIternation bounded polyuomid-time

alternating 'Ihiug m c h e s No and NI which recognize B and B, respeaively. Now we construct a semi-

deterministic alternating W g mackine which recognizes A as foRows: on input z, simulate M on input

z except for oracle queries; whenever M queries yi, guess its oracle answer mrs(yi) and store d u e s yi and

mrs(~i); then nniversaIIy choose [logp(n)j bits c, and if c is the ith element of C ~ ~ S ~ " ) ) , then erase aii

symbols except yi and m(yi), and simulate No on yj ifm(yi) = 0, or else, simulate Nl on gi.

Conversely7 take a polynomial-time semideterministic alternating Thing machine 116. We build a de-

terministic oracle machine M' and an oracle set B such that L(M) = L(M, B). Step by step, we decide

a computation path of M on input x which leads to an accepting configmation of M on x if one exists.

Assume that h, - . ,6, are determined, and M' is at the (m + 1)th node c which has two children a, Q . To decide which chiId node to choose, we follow the pfocedum fonow the leftmost path until M enters a

universal state, and then choose the path IabeIed with the (m + 1)th string; letting p be a label of this path,

query a string {z ,p) ; if the oracle auswers then Iet k+l = 1, and otherwise? let = 0; then choose

%+, and go into the next node.

Oracle set 3 is d&ed by the f o I I d g algotithm: on input (2, p), simulate M on x and det ' . ' , t idy

foflow a computation path IabeIed with p and then simulate the rest d the computation tree T. Since T has

(k - 1)-alternation, B beIongs to ATIMEE(k - 1,n0(')) = Eft-,. Clearly, we have L(M) = L(Mt, B). D

The relationship between the polynomial-time hierarchy and the hear exponential-time alternation hi-

emdlyissmnmarizedasthefollowing~emma

Proof. Use Book's tdlyencoding techniqge [14]. For a set A, define a tally part of A as Tally(,4) = {On I the binary repeesentation of n is of the form lw and w f A). It is not d i fEdt to see that -4 f

-(kt $'("I) if and only ifTalIy(A) E ATIMEX(k, n0(l)). Hence, A E if and onIp if Tdly(A) f X: . A similar ecpivahce relation also holds between At, and Aft. The lemma, therefore, follows h these

characterizatim n

Definition 2.5.9 (Low and High Hierarchies within NP) [89] Let n 2 1.

CHAPTER 2. FOUNDATIONS OF CCOMPMONAL COMPLEXITY THEORY 35

2.5.4 Polpornid-Time Reduu'bilities

PoIynod-time redua'bilities play a very important role in computational comp1exity theory We briefly

sketch such reducibiities.

A set D is d e d pd--time ~~~lfy-m reducible (pxt-edncible, for short) to a set E, denoted by

D Ey if there exists a frmction f in FP such that, for all x, x E D if and only.if f (z) E E. This function f is Caned a ( p o I ~ - t i m e mmgpone) redudion and is said to reduce D to E. Mermore ,

if f is one-one, then we say that D is p o I ~ - t i m e 1-1 dur5ble (pl-reducible, for short) to E. A set

D is polynomial-he M g dtu5bIe @T-reducible, for short) to E, denoted by D 5; Ey if there exists

a deterministic polynomial-time Turiug machine M such that D = L(My E). A set D is p f ~ d - t i m e

hdh-table d d f e (ptt-reduciibIe, for short) to ET denoted by D <L E, if there is a polynomial-time orade

ntring machine M which with orade E makes nonadaptive queries such that D = L(MT E). A set D is poIgmmiaLtime many-one (lb+ng, hth-table, resp.) cotnpw pm-compIete, for short,

@T-comp1ete, ptt-complete, resp., for short) for a dass C if D E C and every set m C is pm-reducible

@-T-reduaibIe, ptt-redudbIe, m p . ) to D. For a oomplexitp dass C, we simply say that D is C-complete if

D is pm-compIete for C.

One of the most usefnl NP-complete problems is the bounded h&g problem, BHP, d&ed as follows:

assmning that {Milia is an effective enmneration of aIl nondeterministic polynomial-time 'Ihring machines,

Iet

BHP = {(~i,z,l~) accepts z within time n ).

We quickly sketch the proof that BHP is NP-compIete. To see that BHP € NP, it is enough to check the

following aIgorithm=

begin nondeterministic algorithm for BHP

Y if y is not of the fonn (%,z, In) then re-

(Now assmne that y = (si,z, 1"))

&date nondetem&&idy A& on input x for n steps

if Mi does not halt then reject

end.

The rrmning time of this algorithm is bounded by a polynomial in n since each simnlati.on of machine Mi

does not e d n stepsc Hence, BHP E NP. Next we show that BHP is NP-hard. For my NP set A, take

awn& ~ ' b t i c ~ ~ e M w h i & ~ A I n p o l ~ O m i a l t i m e , A l s o + ~ a s t r i d y m Q e a S f n g

CZIAPTER 2. FOUNDATIONS OF C O M P ~ O N A L COIUPLEXITY THEORY 36

poIynod p such that TimeM(x) 5 p(lzf) for a11 sQbgs X. Let i be an index such that L(M) = L(Mi).

Now let us d e h e f as f(z) = (&,z, ldCI)). The function f reduces A to BHP and is dearly oneone,

phonest, increasing? length-preserping, and P-computabk

Notice from the above proof that every NP set is pl-reducibIe to BHP by phonest, monotone, length-

presenring reductions.

Another typical example of NP-complete sets is the s- problem SAT that is defined as

SAT = ((0 I F is a satisfiable formnla ),

where (F) denotes an appropriate b i i encoding of a formula F. There is a phonest, onmne reduction

h m BEP to SAT; therefore, SAT is also NP-complete-

The TMng closure (many-one cto- truth-tdle &sum, resp.) of a class C is the coIIection of sets

which are polynomial-the 'Ihring (many-one, truth-table, resp.) redua'ble to some sets in C.

kmm 2.5.10 Eaeq set in A?, k > 0, is p-m-du& to some set in A;. That is, the mrmyone

&sum of A; is d y A?.

Roof. By a padding argument. Assume that A is a set in A?. There exist a polynomial p and a

semidaerministic d ~ ~ g 'Ihriog znacbe M which, on inpat z, recognizes A in time 2p(Irl) with k-

alternations. Let N be another marhine that, on input x of the form zOIAlrl), simulates M on input z.

By dehition, it fonons that ~ ( z 0 l ~ ( b l ) ) = M(z ) for dl z. It is important to notice that N is 2°(n)-thoe

bounded. For the desired reduction, dehe f (z) = d l d w ) for all z.

Two sets A and B are called p o l p m i d f y isomorphic (p'imorphic, for short) if there exists a P- computabIe, pinvertiiIe bijection f which reduces A to B. This reduction f is called a potynomicrl-time

isomo~phism (piirnorpht'sm_ for short). Berman and &atmanis [I01 raised the question of whether NP-

complete sets are aIl pisomorpbic This is known as the "isomorphism conjecture."

A set S is P-printcrbk if there is a P-computab1e fmtction f such that f (On) outputs the list of aIl strings

in S n F. It is know that a set A is P-printable if and only if A is pisomorphic to some tdly set [I]. It is

eaq to see that if P = NP, then dl sparse in P are P-printable 111. A set S is ( T m ) self-reddk if there exists a deterministic o d e Thing machine M such that:

(i) S = L(M, S); and

(ii) On every input z, M queries onlg strings whose length is d e r than 121.

For scampIe, SAT and BEP are both &-reducibIe

2.55 Complexity Cores

We shaII review a notion d a ~ l c p & j t y cars and their existence shown by Book and Dtt [16].

CaAPTER 2 FOUNDATIONS OF COMPUTATIONAL COMPLEXITY THEORY 37

D-on 2.5.11 (Complexity Cores) Let A be a set and let C be a dass of sets. An infinite set H

isa complezity cote (or hardcurt!) for4 rrrithrespedtoCif, forevery set CinC,ifCS.4, then C n E is

finite A compIexity core H is d e d pwper if H E R

Lemma 25.12 [IS] Let C be a m a m i d y entanarrble c h of nmrsbe sets. If C is closed under

finite ufilUfilon andjhde midion, then myinfinitt recurJive set A not inC has an injinite recutslae, proper

cmnpkxify core for A with respeEt to C.

Proof. Assume that A is an infinite recursive set not in C. S i C is recursively enumerable, alI subsets

of A that are in C can be effectively enumerated as {Co,C,, . . .). For each k E N, let DL = &, Ci. Note

that Dj Dj+t for every j . Now let D = Ujlo Dj . I f A - D is infinite, then A - D is an hibite proper

complexity core for -4 since any subset C of A that is m C is a subset of D, and thus (AD) n C = 0. Now let us assume that A - D is finite. Them are h h i t d y many k satisfying that Dk # Dk+1 since, otherwise,

C is closed tmdet hite union, and thus A is in C, a contradiction. For each k with Dk # Dk+l, take the

element ak that is the m b h d in Owl - DL. Let H = {ak I DL # Dk+l)- This E is dearly infinite and

aIso a proper complexity core for -4- t l

2.6 One-way Functions

This section will define one-wuy j h d h s . A one-way frmction is a function which is computed easily but

whose inverse is hard to compute. We shall introduce the new notions of newly-RP and nearfvBPP sets

and show that if one-way ftmctiom exist, then an NP sets are nearly-BPP using h h fvndiaz t&gue

2.6.1 Hash Fnnctions

We shall introduce hash fmdow as a useful tool in the discussion of randomized algorithms.

For n, m E N (n < m), let H,, denote the f d y of pairwise independent unioersd hash fwreCionr from

E" to Cm which is defined as follows: a hash fnsction h m H,, is of the form h = (M, b), where M is an

m by n bii matrix and b is a bit vector, and takes its value as h(z) = Mz @ b. Hence, the set Hw, can

be identified with the set of all m by i a + L over {0, I), and each has6 frmction h is encoded into a

string of length m(n + 1). Note that IlH-[f = 2m(n+1)-

Fu n and c and asmme i I n and UX[l> 0- We say that a function h h m F to F* i 4 M n g d k

CHAPTER 2. FOUNDUND4TIONS OF C0MPUZKEON.L COMPLEXITY TEEORY

z on X if h(z)&- # h(w)&* for aII w E X - {x) ; otherwise, h i-indictinguihes x on X .

Proof. 1) Fi x and i, and let p i , = R h [ h i-distingnishes x on X I h E H,,+& Then,

The second ineqdity is shown as follows. In the CalcnIation, we omit the term 4 E H&,,*-"

In computational cumpIexity theory, there 9e sewd dehiti01ls d mtc-woy fonmbrrr A frmction f is p o % i n m n i o n y ~ k @ i ~ I e , f ~ & a t ) ifthereisafrmctionginFPsarh thatgof(2) = x h r

CHAPTER 2. FOUNDATIONS OF CW-ON& COMPLEXITY TKEORY 39

all x, whereas f is w d y p-inoerEible if there is a P-computable function g such that f 0 g(z) = z for all

z E ran(f). Note that if f is weakly pbvertiiIe, then we can determine whether z E ran(f) by checking if

f 0 g(x) = 5.

A ( w d y ) one-way function is a oneone, phonest, P-computabIe function on C' whose inverse is not

computable m polynomial-time (d [80,45,4]). It is shown in [33,53] that one-way functions exist if and

onlJrifP#UP.

In cryptography, slightly di&rent one-way functions are used. We need only uniform one-wayness m

this thesis. A (my') sftwrg one-wq j i m d h is a ftlnction f such that f is P-computabIe and, for all

randomized 'Ihring machines M working in polynomial-the, the function

is negiigi'b1c Hhtad, hpagliazzo, Levin, and Luby [40] showed a dose relationship between the existence

of strong oneway functions and that of pseudwandom (number) generators.

Let us introduce RP-like and BPP-like sets, d e d d p R P and ndvBPP sets, respectively, which

look Iike onesided and -sided, bounded-error probabilistic sets on most instances.

Dehition 2-63 (NearIy-BPP Sets and Nearly-RP Sets) [I191

1. A set A is dg-BPP it, for every po1ynomial p, there exist a set S and a po1ynomial-time randomized

Th ing machine M such that, for each 2,

(i) z E C' - S implies PrM[M(z) # A(z)] 5 $; aud

(ii) R . [ z ~ S [ x ~ ~ ] < ~ f o r ~ o s t d l n .

2. A set A is nearly-RP & for every poIpomial p, there exist a set S and a po1ynomid-time randomized

lhringrnachine M such that,fmeachx,

(i) z E A - S implies RM[M(z) # A(=)] 5 f ;

(ii) x E a - S implies R M [ M ( x ) # A@)] = 9 and

(iii) ~ [ z ~ S ~ z ~ ~ ] < ~ f o t ~ o s t d n -

We can amp@ the success probability of M on string mpds from A - S by repeating its computations

at random. A new Thing machine N is defined as fotloaa on input z (n = [xi), repeatedly nm M on input

z mdependdy p(n) mang times, and accept x if and ody if M(z) = 1 for some trial Consider z in A - S.

Then, the enor pmbabiITtp that N(x) = 0 is at most ( f ) ~ ( ~ ) . Kence, RN[N(z) = A(z)lz 1 - 2*"). On

the other hand, if z E x- S, then RN[N(z) # A(d] = 0. Without loss ofgeneralitg7 we can M e . aomme

that the length of alI nondeterministic computation paths of N cm t is d y p(Iz1) for some plynomid p.

CIearlyI h m the dehition, RP (BPP, resp.) is properly codained in the dass of nearly-RP (neady-

BPP, resp.) sets

Here we show that the assumption that every NP set is nearly-FtP implies that no strong one-way

functions exist.

Proof. Assume that every NP set is nearly-BPP and a strong one-way function exists. By [9], there

exists a length-preSeroIng strong oneway frmction which b oneone on at least f i ddemm in in br each

n, where p is an adequate mcreaSing polynomid Let f be such a function. Let D = (x 1 11 f-l(x)Il = I). We then have [ID r l Fll> & for almost dl n.

Denote by ST the ith string of c*(") (in particular, s t = p-(")). Consider the folIowing sets (T,),: for exh x,

T r = { z s ~ ( 3 z E ~I=Ir (z )=zh the ( i+ l ) - thb i to f z i s l ,OI i<n I). Note that T, n T, = 0 des s z = y. Let A = U,T,. Clearly, A is in irr- By our assumption, for

the polynomial 4n2p(n), there is a set S and a po1ynomiaEtime randomized Turing machine M satishling

conditions (i)-(in) of De6nition 2.6.3(1) wi th [ISII < & for almost dl n. In parti&

Hence, we have:

We assume that, for z 4 S, RM[M(z) = A(%)] 2 I - 244. We then d& the randomized Turing

machine N as foIlows:

begin randomized algorithm for N

inpd z (say, n = 14) let z s X

fari=Oton-1

if M(+) = 1 then let z s zl else let z = z0

end-far

CEAPTER 2- FOWATIONS OF COMPUTATIONAL COMPLEXITY THEORY

Figure 29: Predicted indusion reIationsbips

Thus, we obtain -V(N(z) ) = z] > 112. Using this inequality, the probabii that N computes the

iwerse off is bounded by

This contradicts the one-amguess off.

Figare 22 iaastrates the inclusion relationships we predict,

As can be seen, nearly-RP sets and nearly-BPP sets are related to the dms?y of sets.

* n 2 6 A set S is C-f ( n ) - h e if there exi& a set B m C such that 11(,4AB) n PI1 <_ f (n) for

ahmostalln

(XUlTER 2 FOUNDATlONS OF C O ~ ~ O N A L COMPLEXITY TIIEORY 42

I;emmo 26.6 Let f kofvndionmNstrchfirabf(n) Ew(logn). FwCE {RP,BPP), euery~-2~-f(")-

dose set is nearlyC.

Proof, We rh$l estabIish the lemma for C = BPP. Sice f (n) E ~(Iogn) , we have fim,,,, = 0

for all numb ere>^. ~implierttut&--KxI&=~fororeryconstantc>~. ~ence,2-f(") G O ( & )

for any polynomial p.

Let A be ~~P-2~-f (")c lose- There exist a polynomial-time probab i ic W g machine M and a set

B such that II(AAB) n Fll 5 2n-f(n) for almost all n. In other words, Pr,[rl(z) # B(2) 1 2 E En] 5 2-f (n).

We d&e S = (z I A(z) # B(z)). Then, we have Rn[z E Sj < 2-f("). By the above calcalation, for

any polynomial p, PG[X E I for almost aIl n. This yieIds the desired consequence. 0

2.7 Relevant Theories

In this section, we shall discuss several important and a branches of computationaI complexity theory-

2.7.1 Feasi'ble Real Numbers

A real number is viewed as an infinite sequence of dyadic rational numbers which converges to it. If the

convergence rate is fast enough and the nth dement of the c~~~vergence sequence is &ectively constmctiile,

we can obtain a good approximation scheme for the real nnmber m questior~. KO and fiedman [w called

such real numbers comprrtable d numben and initiated polynomial analysis of the real numbers based on

real nnmbers computable in polpornid time.

D M o n 2.7.1 (Compntable Real Numbers) [55] Let t be a function on N A real number r is

t-time wmpdabk (t-space ctm@abIc, resp.) if thete exists a detcr-minisr'c Tnring machine M which, on

input In, produces dyadic rational nomber d, in time t(n) (using space t(n), resp.) satisfying the condition

k - dn( 5 2-". We cdl the sequence a mntrergm scpeme fbr r.

D M o n 2.7.2 (ConrpatabIe Sequence) A sequence { ~ n ) ~ ~ of real ntrmbers is polynomial-time

computubk (P-computable, for short) if there exists a polynomial p and a deterministic Thing machine M which, on input (In, It), produces a dyadic ratio& rmmber d in time p(n + k) satisfping Id - znl I 2-k-

In polynomial analysis, frmaions map from gl to Be Although these real functions are intriguing, frmctions

we are now interested only m fanaim that are "discre@ mappings h m C to the tmit intend [0,1]. Here

wemodifytheregulard~offeast'bIedfnnctionstosnitotlrsetting-

Dehition 2.73 (Computable ReaEValned F h c t h s ) A function f h m C to !P is d e d t-time

c o m p a t a b k ( t - s p a c e ~ r r s p . ) i f t 6 e t e ~ a d ' -'-'c'hingmacbineMwhichworlcsintime

CHAPTER 2 FOUNDATIONS OF COMPUTATION& COMPLEXITY THEORY

t (using space t, resp.) that 11\8(2,1i) - f (z)l 5 2-' for all i E N and all z E Em.

Hence.

Proof. Since { ~ n ) ~ ~ ~ is computabIe, there is a deterministic Zhting m d h e M 1% - M(ln, 0') 1 < z4 fix all n, i E N. Let N(ln) = ~( ld"+') , on+'). More precisely? N is a d-c 'hhg

marrhine d&ed as bedaar:

CBAPTEa 2. FOUNDATIONS OF COMPUTATIONAL COMPLEXITY THEORY

compute p(n + 1) (needing O@(n + I) + n) steps)

(*I simulate M on input (Mn+'), on+l)

end.

Ifp(n) < n, then we use the convention that whenever M tries to check the i-th bit of the input, it writes

down i in b i i on a checking tape, and then it retrieves the answer automatidy.

Then, the nmaing time of N on input 1" needs at most

c - ( ~ i e ~ ( l ~ ~ + ' ) , on+') + n) I c (t@(n + I) + n + I) + n)

for almost all n, where c is an appropriate positive constant. We next daim that Ir - N(ln)l I rn. This

is seen as f0Uows:

Therefore, r is computable in time O(n + t@(n + I) + n + I)).

The interested reader may refer to [a].

2.7.2 Kohogorov Complexity

In this thesis, we ase Cime-botmdcd generrJizod Kolrnqmw complex@ give^ by Hart * [38]. Intdidy,

the Kolmogomv complexity of a W e string is the length of the shortest program that will generate the

h g .

To dehe Kalmogorov complexity, we need a notion of lnrirrersd %+ng m c b e j that can simniate the

behavior of any other M g machines. We assume that {MiliEN is au & d v e enumeration of all lhring

machks. Now mnsider the follawing machine U: on input (i,x), U simnlates the ith machine Mi on input

2.

-4 code is a prefix-code or instadaneotts code if the set of code words is prefix-free, ie, no code word is

a p& of another code word. A prek-mde is d e d self-delimiEing if there is a Turing machine which (i)

deddes whether a given word is a code word and (ii) computes the decoding frmction.

Letusfixsncharmiversa'IhringmachineU.

&hition 2.7.6 (Kohogomv CompIesity Sets) [38] A Cimedowrded Kobnogm compkdy set

KTb(n),t(n)] is the set of z such that, far some string y of length at most g(lzl), U on input y

outpntszmtimet(l~). Finafixed~tringz~aEime-boundedcondrkoncrl - - Kohogmoo complQiPy set reloPioe

to z, KTb(n), t(n)jz], is simihiy de6ned but using U((g, z)) = z instead of U(y) = z m the above defbition.

CTlNTER 2 F0ITM)ATIONS OF COMPUTATIONAL COMPLEXITY THEORY 45

Proof. For an output string z , we need to check at most all machines Mi coded by strings i of length

g([z1), and, at each i , we simuIate the machine Mi within t(lz[) steps. 0

For more discussion about KoImogorov complexity, the reader may refer to [62].

2.7.3 Resource-Bounded Measure

in 1990, Lutz [a] developed resrmm-bounded meMwe theory. The foUowing terms mainly folIow [TO].

A function d h m C' to Igt is a martingale if

for all strings w. For every martingale d, we aIwaps have d(w) 5 2Iw1 - d(X) for aIl w. Recall that si is

the ith string in E'. A martingale d succeeds on a set A if h k - snp,,, - d(A[O ... n]) = oo, where AIO ..a]

denotes the string z = z o q - - *zn satisfying zi = xA(si). Let SOD[dj be a collection of all sets on which the

martingale d succeeds.

The notion of pmeosure captures in a way the "topologid size of a class.

Dehition 2.7.8 (P-measure) cf. [a, 701 A complexity class C has p m e m 0 if there exisfsi a

martingale d which is compubbIe by a deterministic polynomial-time Turing machine such that C G Sm[dj. A complexity Jass C ha pmepna 1 if the complement CO-C has pmeasure 0.

The class E has pmeasure I, but P has pmeasure 0. We sometimes informally call a dass smut2 if it has

p -meam 0. According to this terminology, the dass P is smaIl, while E is not small. One of the in-

open question is whether N P is d If P = NP, then obviody N P is small; hence, smaIIness of N P

wodd foIIow from a collapse of NP down to P. A I-dimaui4ncf mmtingate Jystem (I-MS, for short) d is a function born N x C to R? such that 4 is

a martingale for each k E N, where dc(w) = d(k, w). A set X is a punion of tlie pmeosrae O sets (Xi)iE~

if (i) X = UjqN Xi, and (ii) there exists a polynomia-time computable 1-MS d such that Xi S00 [dj] for

- i E N

Jkmma 27.9 [MI If X is a punion of the pmeamre 0 sets, then X has pmeosure 0.

WefmthersaythatChas meruureOinE (meastae 1 inE,resp.) ifCnEhasprneasure0 (p-measure

1, resp-). It is dear from the defhrition that i f C has pmeasnre 0 and 1, then C has measure 0 and 1 in E,

respectiveIy- As an example, we note that, for a fixed positive constant c, the class DTIME(O(2C")) has

measure 0 m E [ 7 ' - Another mid example of a dass which has measure 0 m E is the coIIection of all

pm-cornpkte sets for E (see PI). WeshdIgiveanhqmtantexampIeofpmeasureOsets immtrnesetsand6i-immuneJets

Definition 27.l0 (Immane Sets and Bi-fmmnne Sets) Let C be any complexity Jass. A set S

CKAPTER 2 FOUNDATIONS OF COMPUllATIONAL COMPLEXITY THEORY 46

is called C-immune if S is infinite and S ha. no infinite snbsets in C. A set S is C-bi-immune if S and its

complement are both C-immune.

Pmpodim 27-22 [a] Let c > 0. The chPs of aU DTIME(O(Fn))-bi-immune sets has pmeamz 1.

Hence, any c h to&& eontainr no DTIME(0(Yn))-bi-immune sets has p m m 0.

ProoE Let C be the dass of non DTIME(0(2Cn)>b~immurte sets, and to obtain the desired result, we shall show that C has pmeasme 0, because this obviodp impfies that the dass of aII DTIME(O(Fn))-' m e

setshaspmeasnrel. T& a nniverd set A in E for DTIME(O(Fn)), namely* DTIME(O(Fn)) = (A, I i E N), where

Ai = (I I {qi) E A) for each i. we decompose C into bhitdy-many sabdases {Ym)mEN as iotiow~:

We dehe the 1-MS d as follows:

d(m, z) = 2 d h w ) ifslwl E Al(m+rl/21, and i f m (mod 2),

i f s l ~ l E A~rn+1)/2j and b m (mod 2),

where w is the string satisfying r = wb for some b € (0,l).

To compute d(m,z), w e shonld check if sl,,,~ E At(m+tlpJ; however, since slwl is of length Ilog(llol)

(= llog(lwl + I)])* the computation time for checking if slWl E AL(m+l)121 takes at most d 2~-wl~l) 5 d - ([wI+ 1)E, where c' is a constant. Thus, d is P-computable.

It &ces to show by Lemma 2 7 3 that C is a punion a£ the pmeasure 0 sets {Ym)rnEn. Let us see that

Ym C SOD[&] for each rn E & where &(w) = d(m,w). F i h d e r m E N such that m is odd I£ A,,,

is iinhe, then we clearly have Ym G SOO[L]. We then assmne that A, is infinite, Take any set B E Y,.

Note that AL(mC11/21 B since B E Y,. By our dehiion, we have &(B[O..nl) = &(B[O-n - I]) if

Sn E A~(m+1)/2~ t and otherwise, d,,,(B[Od = &(B[OA - ID- Smce {n [ s, E -4L(m+1)/q) is hhite, Iimsupn- &(B[O.al) = oo. Thus, we have B E S00[&]. A similar argument works for the case where m

iseven. 0

Chapter 3

General Theory of Average Case

Complexity

3.1 Introduction

Avefageca~e analyses have been performed to measure the complexity of algorithms and to obtain a better

understanding of the behavior of aIgorithms when input instances are given with some probab*. In this

type of adysis, we should take into consideration that instances of a particular aIgorithm occur with some

probability. In contrast to the approach of b m worst-case anaIpsis, a "problem" here is a pair consisting of a

set of hdmces and an input d k t d u t h which designates the probability of each hstance. These probIems

are called d i&ih&d problem, madonaked problem, or tondom p b k m

Classical average-case analysis rises the expected running time or tape squares m use over an input

instances d the same length, Atthough the notion is simple and intnitioe, it is not a basis for a consistent

and coherent theory of a- mmp1ait.y- Levin's theory of average NP-compIeteness uses instead the

notion of poEynornial on the a o q e and also plynomicl-the mmry-one rcdt&i& among distributiond

decision problems with crucial conditio~m, the d e d dominoEton cmddmw . - for reduction hctiotls This

c3onstraint is essentid in the theory to make the r e d u c i i ~ transitZve and to make the average polynomial-

time cornpaable cIass dosed under the reductions In this chapter, we shaIl review M ' s theory and farther

cultivate a generat hmework of amrageme c o m p l e theorp; we sba[l focus mainIy on the notiom of

po- on peaage and plylomial darination nhticms on which "domination conditions" rdy.

Levin's notion of frmctions being porynarial on p-attaoge has been expanded into random frmctiolls to

cope with the average-case analysis of randomized algorithms.

Section 3% wiU begin with the f d defiinition of disbibufioru (or distn'bation Functions) and (p6-

abilitq] density frvnclions (or probability distributions)- For practicaI reasons, we s h d introdue semi- Wddicnu by e b h t h g the condition that distributions muverge to 1 as input - get larger- Of parthhimportance is the ~ d i s f r i b o p i o n t h a t asiptoeachstringtfte probabilitpofthestdag

CHAPTER 3- GENERAL THEORY OF AVERAGE CASE COMPLEXITY 48

being chosen at random.

In Section 3.3, we shall introduce the notion of t on p-werage which generalizes Levin's ori- notion

of "po1ynomiaI on p-amqe? Brie&, a ftmction g is called polyrumdal m p-memge if, for some positive

number k, the expeaed value of 1x1-I - g(z)'/' over all input strhtgs x under distnLbntion p converges. The

reader may notice that the fimctions which are polynomiaI on paverage are in general not time-constrnctible,

or even computable.

Section 3.4 will introduce of the notions of domin&*m &tiom and e r p b h c e relofions among dis-

tributions- These relations are the essential ingredients of Levin's domination conditions for averagecase

reducti011~.

Section 3.5 wil l introduce two types of averagecase compldty dasses, "randomizedn version of worst-case!

complexity classes aad (Caverage time/- boundedm version of worst-case eomplexiw dasses to descri i

the classes which consist of distrr'butiortal decision problems. We shaII introduce two types of notations:

Dist(C, F) for the randomized complexity dasses, and Aver(C, 7) for average complexi~ cIasses. horn the

dgorithmic point of view, we can consider sets of distnibnod decision problems (D,p) whose underlying

problem D is s01ved in time polynomial on paverage- In particuIar, we W introduce awmgeme analogues

of important complexity classes, such as P, NP, BPP, and PSPACE. However, the reader &odd note

that our average rrmning time and tape space are not measured by timeconstructible functions, and as a

result, fundamental relationships among the average compiexity classes are d y diffe~ent h m those

among worst-case complexity classes.

Major C o ~ i . a t i o ~ ~ ~ . This chapter extends Schapire's -tion of "po1ynomiaI on paveragen

to a more generat notion of Yt on p-average." The reader who is familiar with Levin's origind definition

may h d it intetesting that the arguments used in this thesis to show that a given hction is po1ynod on

p-aoeage are vey different from those used m the Iiteratttre. The basis of this chapter comes from Schder

and Yamakami [9q. Lemma 33.15 gives a h p 1 e but &cient condition for proving a given fnnction g to be polynomial on

paverage. The lemma s h m that if g(z) 5 c. nk + ( e / n * ~ ( ~ ) ) * for dl i with I 5 i 5 n and for

alI z E An, then g is poIynomid on p-;werage, where {4)lsisn is a partition of F. Lemma 3.416 shows that, for a random hct ion h with its random-input domain r, if p -F v and

i({(z, s) 1 h(z, s) > q(r - bl))) < l/r, then h becomes poIynomial on paverage

A genedhtion of Schapires d t is presented as m Proposition 33.17- The proposition a c W y shows

thatt for a good set 7 of strictly increasing frmctiom, g is T on p--average if and ody if the -on of

the waIues IzI-' - fl(g(z)). over dl mpat strings 2, a~nverge~ for some fnnction t in 7. Lexnmas 3320 and 33.21 are new d t s , Lemma 3.320 shows a suEcient condition for a random

fimction g to be poIynomial on pamage, whiIe Leuma 33.21 ghm a necessary condition for g-

Lemma 3.322 shows that, provided a function h is po1ynod on paverage and a random bction g

is palpomkd on p x vareqe, if h(f(z)) - ij(f(z)) 2 1 for all x and k V ( x ) I is polynomial cw paverage,

then the com@on k.g(z , h(z), s) is palymmkd an p-wemge-

tXAFlER 3. GENERAL THEORY OF AVERAGE CASE COMPLEXZTY 49

hposition 3.520 shows that, for example, Aw(BPP,F) is weakly PPdesaiptive.

Lemma 3.5.21 shows that Aver(P,F) and Aver(PSPACE,F) are closed under weak description., and

this lemma combimed with Fbposition 3.522 Ieads to CorolIay 3.523. The corollary shows that P # BPP i m p k the separation of Aver(BPP, *) born Aver(P, *) , whereas the two aveqease compIexity classes

coIIapse if P = PP. (These teSul~ foIIow Proposition 3.522, which shows an extensive generahation of a

d t given by Karg and Schder [q.) Our notion of nondeterministic average poIynomial-time is dearly distinct h m what has been d k u s s d

elsewhere. In parti*, whereas P =?NP is a Ion-standing open question, Theorem 3.524 makes the

interesting obsemations that A*, *) # Aver(NP, *).

The new characterizations of Aver(NP, F) and Awr(BPP,F) are presented in Proposition 3.5.30 and

Proposition 3.5.33. In parti&, the Amplification Lemma (Lemma 3.5.31) in our averagase setting, fol-

lowed by Proposition 3.5.33, is fundamental and hds many applications in later chapters. The Amplification

Lemma amplifies the success probabilitp of randomized algorithms which make bounded-errors.

3.2 Distributions and Density Functions

Amqpcase complexity theory handles problems whose input instances occur with specified probabilities.

This section wiIl introduce the basic concepts of d i d d u t h (or &std&m fundions) and (pro&u&ty)

t h d y fmcfhm. Using the terminoIogy m Section 2.2, om sample space R consists of all finite strings over

(0, I), and the o-fidd is (Q, F), where F is the power set of R. We consider a discrete probability measme

on (Q, B') and caII it a (probability) d e h c t i o ~ Intuitively, a density function provides a probabiIity

that instance s occurs. A distriiution, on the other hand, indicates the total probabi i over all instances

d e r than or equal to a give instance. For practical reasons, we also use the notion of s--disttibtrtions.

De5ition 33.1 (Dktributiops) A senzi-distrtm (or semi-c&std&h ftnction) C( is an increasing

fimction 6rom C' to the uuit real i n t a d [O, 11. A disbadion (or distribuk'on ftnrdion) is a semi-distriiution

which converges to I, ie, b- p(z) = 1.

We do not avoid the possibility that the semi-distn'bution p always takes the value 0, ie, i(z) = 0 for

aII 2; we caII such p triaiol. We remark that there is no feasri'fe way to determine in g e n d whether a given

semi-distniution p is is Nevertheless, we are primarily iderested m non-ttivial semidktriitrtiofl~,

Ben-David, Chor, Go1beich, and Luby [9], among otheq Ogen use semi-distnins m their arguments

instead offdl distihutim because their semi-distndmtions can be normalized to fan distniutions without

changing the compIexity. However, this normahation is not always possr'b1e See Section 42 for more

disassiof~

Next we st& define (probabil*) density fimctions which are probability measures on the u-fidd

Definition 3 3 3 (Density Ehnctions) For a disfxiiution p, its associated (prohMty) densi ty fmcth

ji is defined by the probtrbility ji(z) on mpttt string z as f01Iows:

(Note that a density fnnction is often called a ptob&i& distribution in much of the literature (see,

eg., [Ill) and should not be confased with a "distrriution.")

The reader must keep in mind that if p is a distri'b~on, then b(A) = p(X) by our d&tion. AIso note

that p(x) = z,, P(z) holds for aIL strings x. - We have already seen the notation R[q for event E based on a sample space R which consists of finite

or M e seqnences over (O,1). We reserve this notation for events where each bit of a sequence is chosen

at that is, for any property Q on X* and any subset E of R,

For convenience, we use the notation &ji(z) a g(x)" to mean that the probability b(x) on input string

z is proportional to the value g(x) for every z; more precisely, there exists a constant c > 0 such that

b(z) = e - g(z) for aII an. This c is d e d the nomolting constrmt for g.

For a distri'bution p and a set S, let b(S) denote the sum LES b(z). For example, ji(Esn) = p(ln) and

fi(Z*) = hz- ~(4. Recall that distributions are mappings h m the W e set Zg to the unit real intend [0, 11. We also

cope d etlsdles of- input distribuEimu instead of %hiten distributions- Given a distribution p

and a natural number n, the c o d h d &dt&n of p on !P, denoted by p,,, is the function from C" to

[O, 11 that is defined by its density fimction A as foflonsr for each z E I?,

whenever &(F) # 0; otherwise, A(%) is undefined- Srmitarly, let e, be the e r m d i t h d distribuCion of p

on Esn that is defined by its density function kn as foIIows= for each z m Csn,

(XhUTEB 3. GE2WM.L TBEORY OF AVERAGE CASE COMPLEXlTY 51

Definition 3.25 (Flat Distriiutions) [36J A distn'bution p is called pot if there exists a real number

c > 0 such that P(x) 5 rlZla holds for almost dl z- Notationally, FLAT denotes the collection of all flat

distributions*

For a hct ion f on Z', we write pr1 to denote the distribution defined by its probabii firl(z) =

B({z I i(z) = 2))-

Recall that distn'butions defined m Definition 3.2.1 are unary functions. We can also consider mdi-

dimerrsionat d- For a k-dimensional vector (x1,z2,. . . , zk) over E', if a density function fi is defined, then let

Using an &kcthe encoding of k-dimensional vedors into strings (discmsed in Section 2.5), however, we can

always iden* probability &, 22, . . . , xk) with probability ji((zt, x2,. . . , sk)) defined on X'. In this sense,

we do not need to consider all multi-dimensional distniutions.

To simpIify the descriptions of distnitxtions, we use k(z, y) and b(z, y,z) to denote fi((z,y)) and

ji((x, y?z)), tespectively. We dso use a simpEed definition of distributions- For example, the following

schematic definition &ji(si, z, In) = B(si) - i (z) - 2-2n~(n)-1n really means that

One of the most important distriiutions is the positive distniution where each string is chosen "rmi-

f o d f at random. This distriitrtion is d e d stundud- In this thesis, we use the standard distrr'btrtion

P-d, whose dues are dyadic rationai ntrmbers, that are easiIy sampled by the foIIowing randomized

aIgotithm: pick a n a W number n randomly and then pick a stdng of length n randody. To pick a natural number "rmifodf at random, we first define the translation Tr by Tr(X) = A, Tr(00s) = Wr(s),

Tr(l1s) = Wr(s), and Tr(Ols) = Tr(l0s) = # for a string s, where # is the terminal symbol dBerent

from G and 1. We then generate a string of the form s01 or s10 such that Tr(s) is the nth string w i t h

respect to the standard order in F. By a simple estimation, we have fi-&) = f l r l - * ~ l + l ) - 1 7 where

IIog(4 = LWn + 1) 1 *

Fkom this definition, it foIIows that, for every z,

CHXPlER 3. GENERAL THEORY OF AVERAGE CASE COMPLEXfm 53

For the sake of convenience, we sometimes use the same notation UW,. to mean the standard distribution

on (1).*

Deffnition 3.2.7 (Default D ~ i u t i o n ) For a random-input domain r and a semi-distribution p, we

define the default semi-distn'bution jq induced h m p and r as

where Ur is the rarity function d I'.

In the special case that I' is almost total, for every snbset S C r, it holds that

A function g horn E* to W is called d e g e n d e u n k p if b({z g ( z ) = 00)) = 0.

A set T of distriiutions is Josed under k-addition if, for any k semi-distributions, PI, pz, . . . , pk, &om 3, the distniution v defined by b(z) = 2, - k ( x ) belongs to 3-

For two distniutions p and v, let p $ v denote the distri'bution q such that @(x) = $(u) if x = Ou for

some u; i(z) = $6(u) if r = lu fbr some u; and j(x) = 0, othemise. We say that a set F of distn'butions is

&sad&$if,foranytwodistn~ntio~t~pandvmF,p$vkalsoinT.

3.3 A Notion of Easy-on-Average

To establish a amistent and coherent theory of amqeese compIexity theory, kst we must examine a

Etmdameetat notion of computational ''average complexity" of algorithms, after which we wilI Iook at Levink

innovative idea of how to amend the naive definition.

3-3.1 Naive Definition of Average Polynomial Time

In worst-case complexity theory, a problem k t(n)-time annpfezitg if there is an agorithm M computing

the problem which satides the ineqaditp T i e H ( z ) < t(n) fix ahalmost natnrat rmmbers n and dl inputs

z € 27'. A naive notion of a- c o r n p i e , however, is given by the expect& nmuiug time (or space)

of an algorithm over aII instances of the same Iength under a certain conditio~I distniution. More preciseIy7

an algorithm dtf which w d c s m time f teqaires the ineqaality that &,,TimeM(z) b ( x ) 5 t(n) for

amost all n. ThiS natmd formnlation of an a- c o m p I e might seem to be a start on a general

thtmy of mmgecae complexity- Unfkfmatdy, this ddefinitIw has severat dekienaes- We win see sewal

examples beIow- For brevity, we say that a h&b g u"expected poiynomiaI on paveragen if there exists a

c ~ n s t a a t k > ~ s a c h W f o r a l l n ~ & ~ ~ ~ ~ ~ ~ ( z ) - ~ ( z ) ~ n ~ f k [ ~ .

The ikst exampk is teIated to the mhipIicatian offnndions which are expected p o l y n d on p-average-

Example 33.1 [9] Consider a function g which is d&ed on strings of length n as follows: let g(z) be

2" on n inputs 2, bnt let g(x) be n on the other inputs. Now take a conditional distriiution h: L ( x ) = 2-"

for all z of Iengrh n. It is easy to see that the expectation of g, &+I=n g(z) &(z), is at most hn, but the

expectation of029 L1,1,@(4)2 - L(d? =a?&

Even if g is expeckd polynomial on paverage, ga is no longer eqmted polynomia on paverage-

The m n d example deaIs with the composition offunctions which are expected polynomial on paverage.

ExampIe 3.3.2 [9] For simplidtp, let n be ofthe f o r m for some integer m > 0. Consider the foIloaring

machine M: M is computed in steps on 2"12 shkgs of length n and outpttts a string of length mt and, for the rest of the inpats, M reqnires n2 steps to output a string of length n. The expected nmning time

of this machine M is at most 1 + n2 2-"I2, and hence, M runs in expected polynomial time on pavexage.

Now consider another machine N which needs n3 steps. Clearly N runs in pcdynomial time. However? the

composition of two machines, N(M(z)), needs PI4 + n3 - 2-n/2 steps on the average

Hence, this naive notion of 'expected po1ynorniaI on the averagen is not dosed under composittio~

Example 3.3.3 [9] Consider a problem on directed graph G = G(V, E), where V is a set of vertices aud

E is a set of edges. Let ]lVll= n and 11 El1 = m- Assume that there is an algorithm M which works on this

graph G in time t(G), where t(G) = 2" if m < n312; otherwise, t(G) = n2. Suppose the graph G is given

by its (imcident) matdr repfesentation. Siace encodings of graphs are presented by n vertices? the amage

is taken over all graphs G with n nodes. The expeaed rtmning time Ec t(G) - L(G) is at most

On the other hand, suppose the graph is given by its adjacency list. Then, the expected nmning time is not

expected polynomial on pavetage

The naive definition is dependent on the particular encoding of instances of a given probIm,

As we have seen, the naive definition is not suitable for a coherent theory. In 1984, Levin [60] instead

proposed a new measure d "a- polynomial-timen To understand his measare, we again take a dose

Iook at the definition ofwmtxase a m p l e measure. Remn that an algorithm M needs polynomial time

if Tre~(z) 5 1x1' fbr ahnost aIl z Here we trausfii this in- into another form of ineqpdiQc

I%[-' - %neM (z))'/' 1. A nahnd idea then is that the expectation of 1x1-' - TfrmeM(z)l/k over dl strings

of length n, is bounded by 1; namely* - ~ r r n e ~ ( x ) ' / ~ -A(=) 5 1. hrin was motivated by this

ineqnalitg., b r r t h e ~ o n e s t e p f m t h e r a n d t o a k t h i s ~ o v e r dfinife- Hedehedhisamage-

case complexity me- m o x d m p-ammgem by mpirbg that & . T ~ ( Z ) ~ / ~ - B(z) < oo.

CHAPTER 3 GENERAL TEE0R.Y OF AVERAGE CASE CO-WLEXITY 55

S i his &st paper appeared in 1984 at the 16th STOC conference, several criticisms of Luvin's com-

plexity measnre have arisen. One of them is that his formulation does not seem to reflect the polynomiality

of the nmning time on the average. In 1990, E. Mapiire [q wrote a technical report in which he hibiteti

an emerging theory of Levin's awngxaw NPsompleteness and gave an interesting insight into Levin's

central notion of "polynomial on paverage." He gave an equivaIent fbrmuIatim of this notion.

This thesis m d S e s Schapire's characterization of plynomial on p-lraerage and introduces a more general

notion o f t on p-ttuemge for an arbitrary hc t ion t .

Dehition 3.3.4 (t on p-Average) [w Let t be a kction on F and Iet p be a distribution. Let g

b e a f u n c t i o n f r o m E m t o P .

1. The function g is t on p-uuercge if b((x 1 g ( z ) > t(1zl. r))) < I/r for any positive red number r.

2. The function g is 7 on p-ooeroge if there exists a function t E 7 such that g is t on p-average.

Schapii actually used a function with two variables, t(n, r), instead of the form t(n - r) in the above

definition. In most cases, however, there is no practical difference between these two fnnctions. For this

reason, we use the above definition throughout this thesis. (This issue has been thoroughIy studied by Karg

and M e r [a], and the interested reader may ref' to it.)

Notice, from Def'mition 33.4, that if g is t on pawrage, then g is degenerative under ~ r , Moreover, it

immediatdy follows from DeWtion 33.4 that increasing the d u e of r also increases the probability weight

of the set of strings z with the pmpeay that g ( z ) 5 t(lz[ . r), which is 1 - I /r.

Definition 3.3.5 (Polynornial/Lagarithmic on p-Average) A timction g from E' to P is ply-

n o d on p-memge if there exists a polynomial p such that g is p on p-average S i i l y , g is b g a d h i c

on p-memge if g is q on paverage for some Iogarithmic hc t ion q (i, q(z) = cIog z +d for some constants

c,d E R).

The notion of polynomiol on p-auemge was fkst introduced m [q and used m [a, 88,36,9, Ill], whiIe

the notion of [ogmithnde on p-uuerage was defined by [91 and also rssed in [a].

Roof. Let us assmne that g is q on paverage for some logadmk function q. Suppose without loss of

that q(z) = cbgz + d for constants c, d > 0. Notice that q is strictly mtreaSing. Using this fact,

wehave

CaAPTER 3. GENE3UL THEORY OF AVERAGE CBE COMPLEXITY 57

Roof. The proposition follows b m Lemma 33.8 together w i t h the fact that if t is a polynomial, then

Xz.t(&) and Xr.t(2z3) are both polynomials. El

Under some reasonable constraints, we can further replace an input ensemble { j ~ . & ~ ~ - in the above

pq=i t ion by {pa)nEn. Cai and SeIman [19] frst proposed an idea of restricting Levin's notion of pIy- no& on p-merage to obtain a better tim+hierarchy theorem. The foUowing proposition takes a Merent

fomnlation but shows an essential part of a theorem by Cai and SeIman.

Proposition 3.3.10 Let g be a j b d i o n + C to P and let p be a d i d d d o w If h.fi(E*) E

R(n-k) for some integer k > 0, then the /oUotoiflo dateme& are epivdenk

1- g is polynomial on p-wenage.

Roof. Let g be a frmction kom X' to It+. Asnrme that h . j i (Eb) E ~ ( n ' ~ ) for some k > 0.

We 6rst show that (2) implies (1). Assume that there atis& a polynomial p such that, for every n f N aud every r > 0, jiln({x E Z a I g(x) > ~(1x1 -T))) < l/r. Set n = 0, and then we obtain

Thus, g is polynomial on paverage

Next we shall show the other implication. ,4ssmne that g is polynomial on paverage; namely, for some

appropriate polynomial p, ji((z 1 g ( z ) > p(r -1~1))) < l/r h o b for all red numbers r > 0. Wrthout loss of

generality, we may assume that p is iucteasing.

By the assumption for p, we can assnme that, for some constant c > 0, ji(E*) 2 5 holds for all

positive integers n. For simplicity, assume that c is an integer-

Let us define q as q(z) = p ( 2 c - e 3 ) fm & z. W y q is a pol~pomid In particttlar, by the monotonicits

of P, q(r - m) 2 p(m - bd+2)

for dl rmmbers r,m 2 0,

Now let S be thesetofaIlnattaalnambers n for w b i m ~ , is de61&

Thdore, g is polynomial on average with respect to (pa)meN, and this completes the proof. 0

3.3.3 Basic Properties

The notion of "t on paverage: which we have introduced in the previous section, is essential in om average-

case c o r n p 1 ~ theory. In this section, we shaII discuss its fundamental properties which w9 be used M y

in the later chapters.

The Iemma below is the starting point-

PmoE (1) If 0 c r < 1, then i({z [ t(lxI) > t ( [ x l - r ) ) ) 5 1 < 5. .9ssame r 2 1. In this case, the set

{x [ ~ ( [ z I ) > t(lzI - r ) ) is empty since t is inmaiing. Hen- P({z [ t(lx1) > t ( lx l - r))) = 0 < $. Therefore,

XzJ(izl) is t on paverage-

CHAPTER 3. GENERAC THEORY OF A'VERAGE CASE COMPLEXITY

Notice that f i ( {z E I f i ( 4 9 < <) < L Thus, we condude that

since = 1- 0

Lemma 3.3.11(2), for example, enables us to see that if g is polynomial on v-d-average, &then g is

for some constant c 3 0.

La us r e d l &om Chapter 2 that f w m i r ~ g, denoted by f 2 g, if and ody if f (x) 2 g(z) for all x-

Froof. Assume that f is t on paverage Since f (z) 2 g(x) for all z, it follows that { x I f (x ) > c) {z

g(z) > c} for an arbitrary c Then,

Thus, g is t on paverage 0

As we have seen in Subsection 33.1, Levin's notion of "po1ynomiaI on paveragen is superior to the naive

notion of "average polynomial" because the set of frmctions which are polynomial on paverage is dosed

under dgebraic operations, such as +, x, max, and mln. Here we show a more general cIaim that the set of

h E Amme that P({x I f (z) > p(l4 4)) < l/r and fi({z I g(x) > q(kd - r))) < l/r- For the rase

CaAPTER 3. GENERAL THEORY OF AVERAGE C . COMPLEXITY

f x g, take a function s E 7 such that s(z) 2 p(2z) + q(2z) for all z. Then, we have

Similarly, for the case f, take a function s such that s(z) 1 @(z))'; for the case r - f , take s such that

s(t) 1 r - p(z); and for the case f + g, take a fnnction s such that s(z) 2 p(2z) + q(2z). The case max{f, g)

is derived h m the case f f g since f + g majorizes max{ f, 9). SimilarIy7 mini f, g) becomes 7 on paverage

~~ ={f, 9) W o w W f , 9)-

As for "po1ynomial on paverage," we sball show m the followiug lemma that this notion is invariant to

any application of polynomials. This tenma is duabIe in later chapters.

We call an n-ary function f immmhg (or monotone) if ;f(zl,. . . ,zn) I f (a?. . . , y,) whenever Xi I yi

for an i, 1 5 i 5 n.

knZn~t3.3.14 cf,[38] M k z l c n d l e t g ~ , ...,git befvncEiorrPfrwnC t o p a n d l e t p k a k -

q -g pol^ If dl 11,. . . ,gk mr on p-auentge, then k p ( g l (z), . . . ,gk(x)) is also

polynomial on p-auemge

Roof. Assume that, for each i with 1 i 5 k, the function gi is qi on paverage for some polynomial

qi. By dehition, it follows that fi({z I g&) > qi(lxi - r))) < ; for all positive real ntlmbers r. Now let

S(X) = p(a(kt), . . . , qk(kz)) for all z- Since p is baeasbg, p(zl,. . . , zk) > p(yl,. . . , yk) implies that there

exists a nmnber i such that q > gi- So, we have

The fdowing lemma is of a rather simpIer fonn than Lemma 3.3-11(3), but it is helpful to show elsewhere

that a given hction is palyMmial on paverage.

CKAFTER 3. Gl3lWU.L THEORY OF AVERAGE CASE COMPLEXITY 61

x E A,n, g(x) I c - nd + ( ~ / n * j i ( 4 * ) ) ~ . Then, g is pdyMndal em p-ooetcrge.

Roof. Let us assume that g satisfies the above condition. For every z , let g'(x) = rniu{g(z), c- Ixld) and

/ ( x ) = g(x) - g'(x). Siaee Xz.clzld majorizes g', by Lemma 3612, g' can be shown to be polynomial on

Cr-a=mF-

We next show that g" is aIso polynomial on pa- Let no be the m;nimd integer such that, for aII

n 2 for dl i with 1 5 i k(n) and for all x in 4A , /(z) 5 (c/n2ji(4))*. Also let b = max(g"(x) I

Now consider the polynomial p such that p(z) = (2c.~)~ + b. I t is obvious by our definition that g"(X) 5 p(0)

and g" (2) 5 p(l/n(n + 1)ji(4)) for all x E A,*. By Lemma 3.3-1 1(3), g" is polynomial on paverage.

By Lemma 33.13, g' + g" is also polynomial on pawrage S i g = g' + g", the p r d is completed. 0

We call a set 7 suitable if, for every t E 7, every c E N, and every polynomial p, there are ftmctions

st, a, 33 E. 7 such that st(%) 2 t ( a ) , s&) 2 t@(z)) , and q(z) 2 t(r) + c for an z.

Proof. (If - part) Assume that g is t on p,-+verage for some t E 7. Take an incTeaSing polpomial p

such that If (%)Is p(Iz1) for aIl z , and an integer Q such that g(A) 5 g. Moreover, we let function t' E 7 be such that f ( z ) 2 t(3p(z)%) + Q for dl z- Such a fimction exists because of the dcsare property of 7. Note that, for x E P and r 2 1, t (3p( t l -# .Izlr) 2 t ( 3 l f ( ~ ) [ ~ r ) . We can show that g o f is t' on paverage.

For the sake of c o d e n c e , let A* = {z I f (x ) # A).

ForeaChrZ1,

CHAPTER 3. GENERAL THEORY OF AVERAGE CASE COMPLEXITY 62

(Only if - part) Assume that g o f is t on p-average for some t E T. Take an inaeasing polynomial p

such that 1x1 5 p(lf(x)l) for all z because of the phonesty off. Let us take a function t' E 7 such that f(r) 2 t(p(r)z) + q-, for all z, where g is the same constant as defined above. We also take AA as above.

Note that, for x E Ax and r 2 1,

We show below that g is f on prt-averageC

3.3.4 Different Characterization

As stated before, our definition of "pol1ynorniaI on p-average" is motivated by Schapire's definition, and at

ttSt gIance, it appears to be dSerent from Levin's dehition. In the 6Rt part ofthis section, we shaII prove

by a slightIy more general argument that both dehitions are ecpident. For this purpose, we introduce a

ugoodn set of hctions below.

For a set 7 of functions on e, we call T good iE, for every s E 7 and every constant c E N, there exist

fimctions st, dr E 7 such that s(a) d(z) and s(2) < #(x) for dl z. For exampIe, the set of inmasing

po1ynominllE is a good set of convex functio~ts.

Proof. (Only if - part) Wrthout loss of generalitp* assmne that g(A) = 0 and that g is t on paverage

for some t E 7. Then it foIIows that, for any real number r > 0, fi({z ( g(z) > t(? - [ X I ) ) ) < I/*. By

om as~rmnptiort of 7, we can define a strictly increasing fnnction t' E 7 such that t(9) 5 t'(z) for ail z. In

partidart t(? - 121) 5 t((r - 1 ~ 1 ) ~ ) 5 C(r - 1 ~ 1 ) - Hence, this implig that b({z I g(z) > f(lx[ r))) < I/*.

Wdont loss of generality* we may Iet this t' be t m the rest of the proot

This indicates that* for every integer k > 0, p({x E CC [ 1~1-'t-'(g(z)) > 5)) < l/#. Using this

m e t b m d Zlrl v . f i ( z ) * by

The last term is farther bounded above by

(If - part) Conversely, a s m e that t-'@(z))(xI-'fi(~) 5 N for some number N 2 1 aad some strictly increasing function t E 7. Markov's Inequality (Lemma A.1) enables us to get the inequality that

E({z I I~l-~t- '(g(z)) > r 4 ) ) < I / r for any real number r > 0. This yields b( { z 1 g ( z ) > t ( rN[z [ ) } ) < l / t .

Hence, g is )LZ ~ ( N X ) on paverage- S i there is a bction t' E 7 such that t '(z) 2 t (Nz) for dl z, we get

the desired result that g is 7 on paverage. n

For example, the set of poIynomiaLP, 7 = {A# + d) I k,d >_ O), and the set of logarithms, 7 = {Xr.(kIog t + d) I k, d >_ 01, are both good sets of stdctly increasing functions on P (recall our convention:

logz is defined to be 0 whenever 0 (t < I).

In 1995, Irnpaghm [43] pointed out that k in ' s d-on is eqtlivalent to the statement that the

expectation of a ftmction over all stziugs of length 2 n, is bounded above by a poIynomiaI in n.

In the following theorem, we see Sdmpire's [88j aad Impagliazzo's -on [a].

Pmof. Note that if there exis& a stzing z such that g(2) = 00, then we can redehe g(z) = 0 without

changing the three mnditions above- Hence, we may apsame that g(x) < a, for dl z m the fo1Iowing proof,

The eqyhdnce between (1) and (2) comes b m Proposition 33-17. We next &ow that (2) and (3) are

eqnivalent.

We &=st prove that (2) implies (3). Assmne that LMX ~*fi(z) < cn for some real number 6 > 0.

Consider a large integer e > 0 such that &+,>o %(z) c Let no be the minimd number in S, and

~ e ~ t h a t ~ d > ~ * e f t h a t ~ ~ @ ~ ~ ) ~ f - ~hen, forearhn~S,

iXUFT32 3. GENERAL THEORY OF AVERAGE CASE COMPlJXlTY

Conversely, assume (3); namely, C.+l<n - g ( ~ ) ~ iy,(x) - < c - n + d for all n E S. Notice that

since

To simplify the description, set I& = {z E P I g(z)6/3 5 1x1) and El = {z E Z+ 1 g(x)'l3 > Iz/). It is

important to note that if x E &, then g(z)*/3 > 1zI2. Thus, we have

The first term zEEo b(z) is obviody at most I. Let T be the last term, and we shall focus on it below:

3.3.5 Random F'unctions

Randomized atgorithms are a simple tool for solving probIems ~ ' ' on the average. To cope with the nmning

timefspace of randomized algorithms, we need a notion of polynomial on p-auemge for random functions

because the running time/space forms a random function which depends on random seeds produced by the

inner coin &ping process of the randomized algorithm.

Gnrevich [36j and Blass and Gurevich [I21 formdated a notion of polynomial on p-auemge for random

functions b m a random-input domain r by regniring the convergence of the -on of the value

~ Z ~ - ' ~ ( X , S ) ~ ~ over all pairs (2,s) in r, with respea to its probability b(z) and conditional probabihy

Or(=) - 2-1'1- Our fonrmlaton is a modification of their definition, and it wiII be shown to be m e u t to

theirs at the end of this section.

Detlnition U . l B (t on p-Awmqp) Let t be a function on P and let 7 be a set of fimctions on IF-

Let p be a distniution and let I' be a random-input domain. Let g be a random function from I' to P .

CHAPTER 3 GENERrLL THEORY OF AVERAGE CASE COMPLE?UTY 65

1. A random function g is t on p-menye if h ( { ( z , s) I f (z, s) > t(lt1- r))) < l /r for all real numbers

r > 0.

2 A random function f is 7 on p-aoeroge if there exists a function t E 7 such that f is t on paverage.

The above dehition impIies that if a random function f wi th a random-input domain I' is polynomial on

paverage, then the same function f , viewed as a "regnlaf function with two arguments, is p o l y n d on

ppaverage, ie., &(((z, s) 1 f (z, s) > p(l (z, s} J r))) c l/r for some po1ynomial p. The converse, however,

does not hold in g e n d

A simple observation shows that h(&s) 5 t(0) for all random seeds s E r(X) unless f i ( X ) = 0. This is

seen as foIIows. For any positive reaI number r,

As r approaches oo, the probability Pr,[h(& s) > t(0) [ s E r(X)] goes to 0. Thedore, this probability must

be 0.

The I~mma below provides us with a simpIe and &dent method for proving a random function to be

7 on paverage. For the lemma, wle must recalI the dehition of a conditiond expectation of a random

function.

Fix a random-input domain r. Fa a random frmaion f h m I' to P, the onditiond expectation

E,Eg(z, s) I s E I'(z)] for each x is defined by

Roof. Let t E 7 and assume that k&b(z , s) I s E r(z)] is t on paverage; namely, b({z I E,[g(z, s) I s E I:(z)] > t([zl- r))) < I/r for all r > 0. We shall show that &({(z,s) I g(z,s) > t'(lz1- r))) < lfr for

some t' E 7. Fix r 2 1. Smce t is strictly inQeaSingt h({z [ t-'(E,[g(z,s) I s E I'(z)j) > 121 -4?)) < 1/4? < l/r. By

Jensen7s inegualityt since t-' is a strictly increasing concave function, it fdows that

3- GElWRAL 17HEORY OF AVERAGE CASE COWLEXlTY

Now let Dr = { X I E#[ t - ' (g (~ , s ) ) I s E I'(x)] 5 1x1 - 2r). We then obtain

Let us consider the last two terms in the above inequality- The first term is bounded above by 1/2r. The

second, howevert is bounded above by

because of the assumption z E Dr. Markov's Inequality (Lemma AJ) ensures that this term is bounded

above by 1/2r. Thus, h ( { ( z , s) I t-'@(z, s ) ) > 1x1 4?)) < l / r -

Using the faa that t is strictly increasing, we obtain the conclusion that j ir({(z,s) I g(x, s ) > t(lz1 - @)I) < l /r , Now consider another fimction t' 7 such that t (@) P(x) for dl z, and as a result, we

obtain the inequality A ( { ( z , s) [ g(x, s) > t'(lx[ - r ) } ) < l / r - Therefore, g is 7 on paverage.

Note that the lemma shows only a d c i e n t condition, and the converse of the lemma may not hold in

general. However, if we instead use Gin's measme (m Section 2.4), then we can obtain a neeeaarp condition.

For a random hction f from I' to It?", let f be defined as

Roar. For each x, if f (x) > t ( l z 1 - a), then R s V ( x , s) 5 t(lz1- 2r) 1 s s E(x)] 5 1/2, because f (z) is

the minimat d u e k sa-g that R s V ( z , s) 5 k I s E r(x)] > 1/2 Hence, it fo11ows that

Using this tut, wqgn estimate the vdae ji((z I f (x ) > t(Iz[ - 2r))) as foflm:

As a particular ease, if a random fimction f is polynomial on pawrage, then its associated frmction f is also polsnomid on paverage

CHAPTER 3. GENERAL THEORY OF AVERAGE CASE COMPLEXITY 67

When f(x, a) is given as of the f o m TimeM (z; s) for some bounded-error probabilistic Thing macfiine

M, we can also repIace r ( x ) by Q(x) in the above definition of f' and the 1-

Root For simplicity, write v for p x q. Let us assume that a random bction g is polynomial on

waverage. This implies the sjstemae of an increasing polynomial p, such that +({(z, y, s) I g (z , y,s) > pg(r - (1x1 + Iyl)))) < l/r for any positive real number r. SimiIarly, kom the assumption that k.1 f (x)l is

polynomial on paverage, there exists an in- polynomial pj such that k.[ f (x ) [ is pf on paverage.

Let us also assume that h is q on p-average for some increasing po1ynomial q.

We then define the new random-input domain rt as I' = {(x, s) I (z, f (z), s ) E r). To reach the desired

remit, it d c e s t o show that, for some polynomial p, h r ( { ( x , s) 1 g(x, f (z), s) > p(lx1- Y))) < I/+. This

po1ynomial p is of the fo11owing form:

Let us ~IX r, r > 0, and define Dr = {z E E+ I h(f (2)) 5 q(l f (z)l- 3r) If (z)l pf(lzl - 3 4 ) . Notice

that, by the monotonicity of pf, p,, and q, if z E D,, then

It is not diflicult to see that

CHAPTER 3. GENERAL TZEORY OF AVERAGE CASE COMPLEXITY

At the beginaing of this subsection, we mentioned the @valence between our definition of polynmnial

on p-memge and the definition given by Gurevich [%I, and by Blass and Gurevich [12]. We now present

this equivalence in the fo11mhg proposition.

~ d b n 3.3.28 Letr beandom--inputdomainandktg bearundomfunca'onfiomf t o e . The

f o l h i n g statanents are eplrioorenk

Proof. Sirnilat to Theorem 33.18.

3.4 A Notion of Domination

Another c r u d notion introdaced by LRvin (601 is the the notion of polynomial domination among distri-

butions. The dmktim rehCion wiU be used as a part of the dominotion cm?dition which is an important

ingredient of the poIpomial-time reduability among distributional decision problems in Chapter 5. Mu- itivelj speaking, it ensures that if an aigodhn is %st on the averageD for somedistriiution p, then this

algorithm is also "fast on the amage" for all distniufions which are dominated by p- This Win deftne

the polynonrial domirrcrEion reidion and the polpaomid ~ a l e n c e nddion and explore some frmdamenta

properties as prepatation for Chapter 5.

3.4.1 Domination Relations and Equivalence Relations

First we shall give a general definition of domination reIations.

Dehition 3.4.1 (Domination ~ 0 ~ s ) Let p and v be semi-dishiintions-

1- IRttandTbeafarrctionandasetoffnnctiorts~X'to~,tespectioely. Thesemi-distribationv

taominoteJpift(2)-qz) zji[s) h d z E C, a n d v T - ~ p i f t h e r e e x i s t s a f r m c t i d n f ~ 7

suchthatvt 'domhata~

CHAPTER 3. GENERAI; TaEORY OF -4VERAGE CASE COMPLEXITY 69

2. Let t and T be a function and a set of functions on W, respectively. The distribution v merage t-

d o m i d a p if there exists a function t', from C' to , such that t' is t on p-average and v t'dominates

p, and v wetage 7-dominafes p if there exists a function t in 7 such that v average tdominates p

This dehition enables ns to consider polynomial domination relations and averagepolynomial domina-

tion relations.

Deftnition 3.43 (Polynomial D o ~ t i o n ) Let p and v be any two semi- distribution^^

1. The semidistriiution p polynomioUy &minutes (pdominates, for short) v if there exists a pbotmded

function t such that p tdominatg v. For brevity, the notation v SP p expresses that p pdominates

v-

2. The semi-distribution p merage-polyndly dominates (avpdominates, for short) v if there exists a

polynomial t such that p average tdominates v. The notation p 5- v means that p avpdominates

v.

In [36], polynomiaI domination and average-polynomial domination are called dmnincrtion and Merrk

d c m i n d q respectidy.

As previonsly mentioned, poIynomial domination relations were explicitly introduced by Levin 1601 on his

theory of averagecase complexity as a cettain type of reducibility between two distributions which measures

the complexity of these distributions. H this sense, two distriiutions which dominate each other on be

mnsideted to have h o s t the same degree of complexity. We call them equiodent. Equivalence relations

capture the doseness of two distributions and also give rise to an appropriate "apprcndmationn between

them.

We begin by giving a g e n d dehition of equivalence relations.

1. The distriiution p is T-equioolard to v if p 7dominates v and v Tdominates p

2 The distnitttion p is menage Tquimht to v if p averageTdominates v and u averageTdominates

P-

-on 3.4.4 (PoIynomial Equhdence) A semidistriitttion p is polynomiany equioaknt

(pquhb , for short) to another semi-distribution v if p is t-eqnivalent to v for some pbounded function

t; and p is wange-pdyumzioCly e q h k n t (avpequident, for short) to v if p is average t q u i d e n t to u

for some polynomial t- We me the notation p z p v to mean that p is p-quident to to, and the notation

p+ v tomean thatp L a v p q u h b t tov.

CXUKER 3. GENERAI; THEORY OF AVERAGE CASE COMPLEXITY 70

As an example, let us consider FIATT the set of flat distniutions. This set is invariant with respect to

p-equidence relations- ThiP is seen as fonotas. Smce p is p-equivalent to u, it holds thy # 5 ti(=) 6 p(x)fi(z) for m e pbotmded p. Then, we have

Hence, I@(=) - t(z)[ 5 2-14'' for some d > 0, and in conse~uence, p is Bat.

Domination and equivalence relations are transitive.

Proof. (1) Assume that there are bctions rn and p.r in 7 such that m(r) - ji&) 3 iil (z) and gr(x) - ji3 (x) 2 ji2 (2) for aU z. To obtain the desired redt, we set p(x) = = ( z ) - p2 (2). Then, we have

(2) Immediate h m (1). 0

As an immediate consequence of Lemma 3.3.14, we have seen that if f majorizes g and f is po1ynomial

on paverage, then g is also polynomid on p-a- Wow we shaIl show that if u avpdominntes p and f

is po1ynomi;ll on v-amage, then f becomes poIynomiaI on paverage.

A class 7 of frmctions is said to be closed under c o t n p o ~ with p o t p m i d s if, for any ftrnction t and

any po1pnomi;rl p with integer d a e n t s , t E 7 i m p b k . t ( p ( z ) ) € 7,

h o t (1) Assume that p 5- v, and h is t on u-average for some h a i o n t E 7. Choose a hction

q which is pdynomid on p-average such that, for dl z E Em, '(2) - q(z) 2 fi(z). Fkst notice that if h is

degenerative under v, then h is also degenerative under p since t ( z ) = 0 implies ji(z) = 0 fix any string z- By

om assmnption, P({z 1 h(z) > t(Iz1 or ) } ) < l/r br alI r > 0. Sime the set 7 is dosed tmder composition with

polynomiaIs, we assmne without loss of genefaIitp that h(A) 5 t(O), and therefii fi((zf h(A) > t(0))) = 0.

Smce q is polynomial on paverage, there exists a poiynomial p such that b((z I q(z) > p(lz1- r ) ) ) < l / r for

d r > O .

CELAPTER 3. GENEWL THEORY OF AKEMGE CASE COMPLZXWY n

Let p,, and v, denote the conditional probability of strings of length n of p and v, respectiveIy. Note

that if q(x) 5 p(lzl. r) for a string z of length n, then &[x) 5 p(n . r) - C(z) for alt n E N and r > 0. Now

d&e g as g(x) = t(42 -p(2z)). Smce 7 is dosed under composition with polynomials, g is in 7- We note

that, for all n E N and all r € W , g(n . r) _> t(n - 9 4n2 - p(n + 2r) ) . To complete the proof, we &odd show

that h is g on p-average. This is seen as follows: for every real number r 2 1,

(2) Let us r e d l that v is avpqtident to p if and d y if v avpdominates p and p avgdominates v.

The resnIt thns f011m immediately from (1). 0

Cwkyl?..C.7 LethbeanyfundionfrwnC'toHPtoo. Fordijtri6utionspandv,ifvuup- UP

and h is polynomial on v-marrge, then h is also pEynonu'd on p-avaagc

Corollary 3R7 implies that the avp-equivaIence relations preserve the notion of polynomial on average.

This faa motivates IIS to introduce the two new notions of irrch*n'm and epucrlity among sets of disttiiudons.

These notions win be used in later chapters.

Deftnition 3.4.8 Let f i and 3 2 be two sets of distniuti~ns~

I. FI potynomially indudes (pmdttdes, for short) 32, sgmbowy Fl P F2, if every dishibution m 31 is

peqnivaed to some distriintion m 32. SnniIarly, 31 m m g e polynomialy includes (avpindndes, for

short) 7i, spmbolidy 3i P 32, if every distniution in F1 is a q m p h h to some distniution

in 3 2 -

CHAPTER 3. GENERAL TBEORY OF AVERAGE CASE COMPLEXITY

Proof. By Lemma 3-43 and the definitions.

3.4.2 Fnndamental Properties

The domination relations are of importance in average polynomial-time computation. This subseaion wiH

explore properties of domination relations in reIation to functional composition, and prepare the ground for . -

timmdma d m which wilI be introduced and cultivated in Chapter 5.

Let us be@ w i t h several important properties. Recall that, for a function f and a distn'bution p, the

notation prt denotes the default distniution defined by its probability jirt (z) = ji( f" (x)) .

Proof. (1) Asmme that p 5 p v. There is a pbonnded function s from C' to R+ such that f (z) s(z) 2 fi(z). W h u t loss of geueralitp, we w e e that s(z) 2 1 for alI z. Consider an incfeasing polpnomid p such

that s(z) 5 p(lz[) for all 2. Smce f is phonest, there is an increasing polynomial q such that [zI 5 q(l f (=)I) for aII z. Now we Iet t(y) = max(s(z) I f (z) = y). It is dear h m the definition that, for every y, there is

an dement 2, E f -'(y) such that t(y) s(q,). For this z,, we have

Since p o q is again a polynomial, t should be paoUnded- Hence,

(2) Let f be a one-one, phonest, pbounded function on C. Assume that ~ c f ~ SP v. Take a pbotmded

fimction p h C' to HeC such that fi(f-l(y)) I fib) - p(ly1) for aII y. Write z = f"(y). Then, since

y = f (z), we have b(z) < P( f (z)) pV(z)). Let q(z) = p(f(z)). Notice that q is pbotmded because f is

gbonnded. Thus, we get p s p v o f. Conversely, assume that p SP vo f. There is an hereasing polynomiaI p such that b(x) S f ( f (2)) -p(lzl)

for all z- Since f is phonest, lzl 5 s([f (z)l) fix some polynomial s. Write y = f (z). Then, z = f -'(y)

because f is one-one Let q(z) = p o f-I(%), and then q is pbotmded, because

C&APTER 3. GENER4L. THEORY OF AVERAGE CASE COMPLEXITY 73

Lemma 3..#.11 Let f beafunetiononC andletp andv k d i d d M h s . The folloraitrgstatementsanz

fq?Livalentr

Proof. We shall show the implication b m (2) to (1). Assume that p 5P 7 and P(y) 2 ijf I (y) for ail

y. Choose a function p such that p(x) tj(z) 2 P(z) for all z. Without loss of generality? we assume that

p(r) > 1. Slncetj(z) 2 S?

Candy? we shall show that (1) implie (2). Let us m e that t(p) 2 # for d y, where p is a p-bounded function- We may assume p(r) > 1 for all z. Ia @(z) = #$. Obviously, q is a

semidistriiution. We then have

Prod (Only if - patt) A r n e that p 2 q and C(y) 1 iift (y) for all y. ,As we have seen, the last inequaIity can be repIaced by the corresponding eqnality.

WenotetbattheineqnalityP>iifl m(m) mbereplacedbg-theegnation3=ilf17sine83(ya) > fir L (YO) for some go, then C, fib) = & t(y) +fi(yo) > & iiyt (Y) ++ (yo) = 1, a contradiction. Hence, we may assume that fib) = r j f t (y) for all y.

There is a pbarmded frmction p such that p(z) - ij(z) = i(z) hr aIl z. In particular- if p(z) = 0, then

resetp(z) to be 1 withoutchanpinptheeqnationsincefi(z) =O. So, wecana~smnethatp(z) >oforaUx.

Then,

C&APTER 3. GENERAL THEORY OF AVERAGE CASE COMPLEXITY 74

(If - part) Amme that i(y) 2 xref-l(v) for all z, where p is a pbomded ftmction h m E' to I&+. By Lemma 3.4.13, this is equident to p'(y) - b(y) > ,!if! (y) for some p'. Consider p" such that

p"(y)-&(y) = jip (y) for all y. As above, we can assume that p"(z) > 0 for all z. Let pf (z) = p"(f (2)). Then,

Now dehe q as i(z) = fix each z. ClearIy we have pf (y) -i(y) = fib} =h=P(y) = C ~ E ~ - L ( ~ ) p,(t

and w = & -qO) i1(4 = jifr (d- n

Assume that f is one-one. Then, p,' saw v if and d g if tlrete Qists a j b d i o n p which is polynomrynomral

p-wwage P(Y) 2 &f-z(,) # ff, dl Y-

Proof. We shaIl prove both rldmn simnltanedy because the main ditference is the condition on p.

(Only if - part). Asstme that p(y) &(y) 2 fir, (y) for all y. We can assume without loss of generality

that p(y) > 0 for all y. Dehe p'(z) = p( f (2)) for each z. I f p is pbounded, then p' is aIso pbormded since

f is pbotmded and v(z) 1 = b( f (z)) 1 5 q(l f (2) 1) for some polynomial q. If p is polynomial on wt -average,

Lemma 33.16 infers that p' is polynomial on paverage. For y E ran(n,

(If- part). Asmne that t(y) > # hr sane pbonnded function p. Let p'(p) = maxb(r) I z E f "(y)). I f p is pbounded, then p' is a h pbounded, since f is phonest. I fp is poIynomid on paverage,

then using the fact that f is oneone, lemma 33.16 again shows h t p' is polynomial on pp-av- Then,

The foIlowing lemma is a poIynomiat-egnivaence version of Lemma 3-4-13.

Lemma 34-14 Let f k a phoned, p h m d e d m on C', anb let p and v be distributions.

CaAPTER 3. GENERAL THEORY OF AtrERAGE CASE COMPLEXITY 75

Proof. We shall show both claims at once. We assume that q(x) -jirt (2) 1 P(z) and p(z) - 8(2) > f i P (z)

for all z. Let p'(z) = p( f ( z ) ) and let q'(z) = q(f(x)). By the proof of Lemma 3-4-13, we have P(y) I

xzE f-l(r, # for dl y. The other Man foJl0ws similarly~

If q is p-bounded, then q' is also pbounded since f is pbomded- We show that q' is polynomial on

paverage i£ q is polynomid on waverage- Assome that q is SO on paverage for some increasing polynomial

so- By our assumption, p is polynomia on prl-average- Take a polynomial sl w b e s i q the average-

poIynomiaIity d p. Since f is pbotmded, there exists an inmasing polynomial t such that If (x)l 5 t(lx1)

for all z. Let D, = {y I p(y) 5 sl ( ly l - 2r)). Let s (z ) = so(t(z) s1(2t(z) - z ) - 3t(t)* - r ) for all z. Note that,

for z E !8 and r 2 1,

where y = f (z). For r 2 1,

By domination sl(2n7) - 3 ( y ) 2 @,-I (y ) , therefore the last term of the above inegnalities is bounded by

3.4.3 Randomized Doxuination

We also introduce a randomized vetsion of domination reIaticms and avpdomination relations.

DefZnition 3.4.15 (Randomized dominstian) Let p and v be any two semX&riiutions.

(i) the mdom function p is pbo- and

CEAPTE3 3. GENERAL THEORY OF AVERAGE CASE COMPLEXZTY @) p(z , s) - i ( z , s) 1 b ( z , s) for dI pairs (2, s) E r.

2. The semidistribution u is said to randomly menage pdominate p (avrp-dominate, for short), symbol-

ically p v, if there exist a dom-input domain I' and a random fnnction p from I' to N such

that

(i) the random function p is polynomial on paverage; and

&) p ( z , s ) - t ( z , s) 2 h(2, s) for aIl pairs (z, s) E K'.

To emphasize the random-input domain 7, we use the notations p s?' v and p 3- u.

By dehition, it is immediate that if I' is almost total, then p s P v (p sav v, resp.) implies p 5;P y.

(p sMP yy, resp.). We have shown as in CoroIlary 3.4.7 that if v Sa''P p and h is polynomial on u-average,

then h is also polynomial on paoerage. In the following lemma, we shall show an anaiogous resuit for

avrpdomination relation-

Lemma3.&l6 L e t p a n d v b e C m o ~ . LethbearandomfirndionfromI'toP,tahenK'ir

a rmrdom-input domain Asstcme that, fw some poZymmid q, t ( { (z , s ) I h(z,s) > q(r . Izl))) < l / r holds

for any dnmnbmr >O. I f p $ F ' v , then h ispolyrum#'al onp-avemge.

Proof. The proof is similar to that of Lemma 366(1). Assnme that h is q on waverage. Smce p -? v,

there exists a random fnnction p being po1ynomial on paverage such that p(x, s) - t (2, s) 2 br (2, s) for all

pairs (2,s) E r. Let US consider an incrrasing polynomial q' such that p is q' on p - a q .

Wemlkapositiverealmmbert> 1- L a n s d e h e theset D,= ((2,s) E rI X E EChp(z,s) I q'(1.I - 2r)). Let be an integer snch that co 2 max{h(A,s) I s 6 r(X)). Such an integer exists (see the

obseroation made after W o n 3-3-19]. A s the desired polynomial 4, we set 4(t) = q(@ - q'(2z)) +a for

all z. The probability h ( { ( z , s) [ h(x, s) > @(lzl- r))) is bounded by

The &st term is o b v i d y bounded above by 11% Let T, be the second term. To compute Try we note

that if (2, s) E D,, then q(1zI - r) - O(z,s) 2 h(z, 8). Thus, Tr is calculated as foIIowm

3. GENERAL THEORY OF AVERAGE CASE COMpLExITy

3.5 Distributional Decision Problems

In contrast to worst-case complexity theory, ammpase c o r n p i e theory deaIs with not only a problem

D bat also a distri'btttion p of instances. Such a pair (D,p) is cded by many mearchers a diddkthal

p b h [log, 1101, rmuiomiad pn,blenr [w, or mndom pmblem [60]. The distribution p assigns to an

instance the probability of its cxcmence as an input to the probIem D.

Debition 3-5-1 ( D i i ' b u t i o d Decision ProbIems) A &tdtdhd (decision) pmblenr (D,p) is

a pair of a set D of string and a distriiution p, A set of djstriiutiod probIems is called an merclge-me

cmnplezity

This section wil l focus on "naturaln average-case compl1exity cIasses and give their format definitions.

Of particuIar interest are two types of averagecase complexity &sses. FoIIdg their dehitiom, we shall

discrtss g a d separation and collapse d t s among these averagecase cornpleexity dasses.

3.5.1 Average-Case Complexity Classes

S i to worst-case c o m p l e theory* we consider Ycomplacity clasd' of distributional decision

probIems. One natmaI tppe of those classes is the combiition of exhting worst-case compl- ckffes C

and sets 3 of distn'butiom. Such a dars was first introdwed by Levin [60j zs an aperagpcase version of

NP, and kter Ben-David, Chor, GoIdreich, a d Luby [91 invented a general notation {C, 7) to descrii such

elaaes. Here we sI&htly modify their notation and introduce! an apersge-case compIexity ckss Dist(C, F).

De5ition 3.5.2 (Average-Case Comp1exity Cksses) d [Dl La C be a comple~~& dass and 3 be

a dess of distniuti~ns~ Let Dist(C, F) be the set { (D, p) I D E C, p E F).

To simplifp the notation, we devise the foilowing coIlvention: whenever the set of a22 distriiutions is

discrrssed, we use the symbol * (asterisk) as the distribution panmeter- For example, Dist(NP, *) denotes

the conection of all pairs (D, p) such that D is an NP set and p is any distnitxtio~

Here we see two examp1es of distnibutional decision problem in Dist(NP, *).

Example 3.5.3 A graph G = (V, E) is called kdunzbk if there exists a doring c of G (ie, a map

c : V + {0,1,2)) such that, Tor any two distinct vertices u and w, if (y u) E E, then c(u) # c(v). The

hlorcbi I i ty problem ( ~ C O L , M ~ ~ ) i~ d&d a~ ~OIIOWS. kt

3COL = ((G) I G is a gmph which is 3aIorable );

and Iet

imL((q) = ~pg.(~uvu) . rG) where (G) is an appropriate encoding of G. The distrrintion jim~ is best desaibed by the fdOwEng

experimenk randomly choose the number of vertices and then rand- choose edges between pairs of

QXAlTER 3. GENE3?AL THEORY OF AVERAGE CASE COMPLkBZTY

distinct vertices Smce 3COL belongs to NP, the problem (3COL, ~ ~ C O L ) belongs to Dist(NP, *).

-1e 3.5.4 The rcndomizcd 3 sotisfiobility problem (~SAT,RAT) is defined as ~OUOWS:

It is clear that (3SKT,pS~~) belongs to Dist(NP, *).

Another type of amqpaseccmplexity classes is more involved with algorithmic computability in feasible

time on atrerogc First of dl, we shaU give a general dehition of time- and space-bounded on merage for

'Paring &es. Notice that m c and nondeterministic 'Itring machines are special cases of alternating Thing machina We dehe the time and space wmplexity of alternating hving machines in an

avempasesetting.

Dehition 5.5.5 (Time / Space Bounded on Average) Let M be an alternating oracle ' h h g ma-

&he and p a distribution. Let A and S be any sets. Also let t and 7 be a function and a set of functions

on El+, respectively.

1. The &e M wit4 oracle A is t-time borrnded on p-avemge if the function Az.Tune$(x) is t on

paverage, and it is 7-Eime bounded on p-werage if it is t-time bounded on p-average for some t E 7. The machiae M is said to ncogntje S in t-time (T-time, resp.) on p-mmgc if M is t-time (or 7-time,

resp.) bounded on p-average, and S = L(M, A).

2. The marhine M with orade A is t-space bounded on p-mmge if the function ~r~pace&(z) is t on

pawrage, and it is 7-space bmrnded on p-memge if it is t on paverage for some t E 7. The machine

M is said to mognke S ah t-space (T-space, resp.) on p-maage if M is t-space (or 7-space, resp.)

boded on p-average, and S = L(M, A).

We note that, m the tzse that fi(z) = 0, the machine M does not necessdy halt on this particular input

z according to om interpretation, the instance z does not occur. Nevertheless, this is not mciaI

m our theory. SehnIer and Yamakami [9?j, for example, have considered onlp machines which always halt.

For a randomized 'Ihring marrfiine M, we must demand that the random function Xzs.TimeM(z; s) be t

mi-==@-

* . DeBmbm 33.6 (Time Bounded on Average) Let M be a randomized orade Timing machine and padistribntion. Let AandSbeanytwosets. AIsolett andTbeafunctionandaset d frmct io~ l son~ ,

cHAPZW2 3. GENERAL TBEORY OF AVERAGE CASE COMPLEXITY 79

respectively. The machine M with oracle A is Qned t-time bomaded on p-auenuje if the random function

h s . ~ r m e & ( ~ s) is t on p-average, and it is 7-Eime /hm&d on p-uuerage if it is t-time bounded on paverage

for some t E 7. The machine M is said to recognize S in t-bime (T-fime, resp.) on p-merage if M is t-time

(or 7-time, resp.) bounded on paverage, and S = t ( M , A).

We are especially interested in maches which are jwiymnni&time/spue h d e d on p-o0croge.

Dehition 3.5.7 (Polynomial-T0ie/Space Bounded on Average) La M be an oracle Thing

machine with output tapes. Let A be a set and let f be a ha ion . We say that M with oracle A is

polyn-d-time (polynomial-spuce, resp.) bounded on p-aumge if MA is 7-time (7-space, respectiveIy)

bounded on p-average for 7 being the set of all polynomials (Le., any hmctiom of the form of 2, Q. 2, cri E 2). The machine M is said to compute f in polynomial-time ( p ~ l ~ t a - s p a c e , resp.) on p-aoerage if

M is t-time (or 7-time, resp.) bounded on paverage, and f (z) = MA(z) for aII z.

Now we are ready to introduce the second type of averagecase complexity compIexity classes Aver(C, 7).

Dehition 3.5.8 (Average-Case Complexity Classes) d [87] Let t, s be frmctim on N and let 7 and S be sets of functions on N. Also Iet 3 be a class of distributions. Tune and space-bounded mange-case

complexity classes are defined as follows:

Aver@TIME(t), F) is the collection of distn'butional decision problems (D, p) such that p E 3 and D is

computable by some deterministic Turing machine m t-time on paverage- Let Aver@TIME('I-), F ) =

UtE7 A=@TIWt), a- Aver(NTIME(t), F) is the coIIection of distnitrtionaI decision problems {D, p) such that p € F and D is recognized by some nondet -c Thing machine m t-time on paverage tet Aver(Ecl (7-) , F) =

Ulrr A=(NTZME(t) 9 8.

Aver(BPTIME(t),F) is the collection of (D,p) such that p E 3 and D is recognizable by some

botmded-ennr probabilistic %ring machine in t-time on pa mag^ Let Amr(BPTIME(7),F) = UE.rAver(BPTIME(t), F ) Si!uniIarly, Ave@?ME(t),F) is the coIlecti011 of (D,p) such that p E 3

and D is mophb1e by some onesided m r , pmbabilistic lbrhg machine in t-time on paverage

Let A-(ItTfME(7)7 T ) = Ute A = w J m t ) , n* Awr(ATIME(t), F) is the dection of (D, p) such that p E 3 and D = L{M) fbr an al- Tbbg

machine M which is t-time bounded on paverage The class A ~ ~ ~ ( A T I M E ' [ ~ , s), F) is the a ik t ion

of (D, p) such that p € f and D = L(M) for an ahemat& 'ihring marhine M m s-time on paverage

which is t-aI-on bounded, starting with an exbk t td state SmiIarly, 9ver(ATIME?(t, s), 7) Es d e k d by semi-deterministic a k m t b g aIternatiag machines. Let Aver(ATIME(7),7) =

U&ver(=(t),a and MY ~ v e r ( m * V ~ S ) , F ) and ~ v e r ( m V , S ) , ~ ) defined.

CBAPTER 3. G- TaEORY OF AVEXUGE CASE COMF'LEXlTY 80

5. Aver(DSPACE(s), 7) is the collection of (D, p) snch that p E T and D is computable by some deter-

ministic l h h g machine M in s-space on paverage. Let Aver(DSPACE(S),7) =

USES A=@SPAWs), F).

Ben-David, Chor, Goldreich, and Luby [9] use the notation AverDTime(t(n)) instead to denote our

Aver@TIME(t), *) (also denoted by AvDTime(t(n)) in [q). In what follows, we use abbreviations commonly used m worst-case comp1exity theory, such as P (=

DTIME(no('))), PSPACE (= DWACE(~O('))), RP (= RTIME(no('))), etc. (see Section 2.5).

R e d l the randomized 3-colorabii problem (3COL, ~JCOL). It is well known that 3COL is NP-complete

(see eg., [26]). However, Wilf showed that the randomized kolorabilitp problem can be solved by some

detetministic algorithm in time polynomial on ptcot-average Thus, (3COL+, p3COL) belongs to Aver(P, *).

We have introduced Eve categories of fundamental average-case complexity dasses Aver(C,3). We can

extend our dehition to classes which do not fail into those categories. One such extension is the conqkmd

of an average c o m p l e class Aver(C, F).

Dehition 3.5.9 (Complement Classes) For a complexity dass C and a set 7 of distriintions, the

cmp- of Aver(C,a is denoted by Aver(cc4,F) and is d&ed by the coIlection of all ~ i d i o n a I

decision problems (D,p) such that (n, p) belongs to Aver(C, F), where = X* - D.

Another extension is the intersection of two avempxe complexity classes Aver(C1, T ) and Aver(&, 3).

Detlnition 3.5.10 (Intersection Classes) For two complexity dasses CI and C2, and a set 3 of dis-

tributions, the intersection of Aver(Cl, 3) and Aver(C2, F ) is denoted by Aver(C.1 n C?, 7) and is defined by

the dec t ion of all distn'butiod problems (D, p) snch that (D, p) E Aver(C1, F) and (D, p) E Aver(C2,F)-

For example, we can define the fdowing average-case complexity dasses: Aver(c+R.P, F), Aver(-NF, F), A ~ Z P P , T), and AVFZ(NP n ~ ~ N P , F ) .

Remark 3.5.11 The notation "Aver(-, -)" in which we use parentheses referring to funcCions might be

misleading- The reader should resist the temptation to replace a cIass C m the notation "Aver(C,F)" with

another @olart class C because the eqnality C = C in worst-case complexity theory does not always

imply the eclnalitg Aver(C, F) = Aver(Cf, T ) m a- complex& theory-

We shaIl introduce a cormterpaa of the W o n dass FP in worst-case compIexity theory= the mwqe

poCyromiaf-time computable jhctibm Redl that a k a r u d w is a 'lhting machine equipped with an extra

output tape for the prrrpase of computing a fm!ction on C'.

Deanition S&l2 (Average Time ComputabIe -OILS) It a h c t i o n f on C is computed by

a deterministic transducer which is palyno&time bounded on paverage, we say that f is comptde

C&APTER 3. GENERAL TEEORY OF AVERAGE CASE COMPLEX2TY 81

in time p o w mt p-menzge Let Awx@F,T) denote the collection of an pairs (f, p), where p is a

distribution m 3, and f is a function computable m time polynomial on paverage-

We shall take a qnick g h c e at some fundamental reIationships among average complexity classes. By

the definition of t on p-ooercge, the class Dist@TIME(t),F) is dearly included in Aver@TIME(t),F).

S i i indusions obviously hdd for other fnndamental average complexity classes.

Pmp~dio t z 3.5.13 IRf C f {DTIME(t), NTIME(t), BPTIME(t), FtTME(t), DSPACE(t)) for some in-

&fGnction t on N, a d let 3 be a set of distributions. Then, Dist(C,7) E Aver(C,7).

The next lemma foUm immediately h m Lemmas 3.4.6(1) and Corollary 3.4.7.

b o E Since the sets of polynomiaIs, h@tbms, and exponentiaIs are all dosed under co~~position with

polynomials, the lemma immediately follows h m Lemma 3.4.6. 0

SimiIar simulation techniques show basic indusion relationships among average complexity cIasses.

Proof. We use a standard sbdation technique to show the above indusions of a- campIexily

elasseselasses

(1) Obviouslyf d-c lhring machines are viewed as a special case of one-sided error probabilistic

nuiugIM&ines.

(2) It saffices to note that onesided error probabilistic Thing machines are nondetr ' ' tic lhring

&es because of onesided error probabirtJr- (This is not ttue for -sided error probabilistic &a)

(3) Suppose (D,p) is in Aver(W?ME(t),F)- Take a onesided error, probabilistic lhing madim M reoognizing D in time f on paverage Let us consider the simnlation machine N defined as foIIows

CEWI'ER 3. GENEFLU THEORY OF AVERAGE CASE COMPLEXITY

begin randomized algorithm for N

input z

simulate M on input z

ifMacceptsztheaaccept

simulate M on input z

i fM acteptszthenaccept elsereject

end.

Asmme that z E D. The error occurs if both simnlatious of M on z fail to reach any accepting

c o ~ t i o n . Hence, the error probability of N on input z is at most 1/4. Assume to the contrary that

z 4 D. Because of onesided error probabi , there is no error occurring in the computation of N on z.

Therefore, N has bounded-ermr probability-

By definition, the ntnning time of N on z, Time&), is bounded by a hear function in Time&).

Smce M is t-time bounded on paverage, the function Xz.Tnne~(z) is kc(t(z) + I) on p-average for some

positive constant c

(4) Fitst we observe that any botmded-error probabilistic T h i t g machine M can be simulated deter-

ministidy using at most O(Tiie;;d(z)) tape squares. This is seen as follows. We simuIate M on z by

choosing its con&mations one by one (i.a, using breadth-bt search) and simnlating the corresponding

steps. Assume that we visit enough leaves of the computation tree so that we can determine whether M accepts or rejects z. We then quit the simulation and go into an accepting state if M accepts z, or else go

into a rejecting state. This simnlation needs tape space at most c - (Tiie&(z) f 1) for some f k d constant

c>o.

Now suppose that (D, p) is in Aver(BPTIME(t), F). There exists a bonnded-error probabilistic lhring

machine M which recognizes D in time t on p-average. Smce h.TimeM(%) is t on paverage, Lemma 3322

implies that Xz.Tiiek(z) is A2424 on paverage The claim follows from the observation above.

(5) G i a deterministic Thing machine M, which on inpd z nms using space p(z), the rmmber of

possible co&mations given by M on z is 0 ( 2 p ( r ) ) . Hence, the simulation of M on z takes O(Hz)) time

0

Generafly speaking, the function b.Tm!eM(Z) for a nondeterminirtie madine M is not time

constructii1e; therefore, even if t is tirne-a3nstrnctiiIe, the inclusion Aw@TIhE(t),F)

Aver@SPACE(O(t)),F) may not hold. This sheds Iight on the c r u d difference between worst-case com-

plexity and averagecase complexity.

As an exampIe, we shaII demonstrate a closure property of comp1eJdty ckss mder disjoint

union. It is easy to seem w o r s t ~ c o r n p I ~ t h e o r p t h a t if Aand B are in P, then A@ B is also m P.

Intheaperagecasesetting, wemustaonsiderthedistribntionsasdasthesetsofstrin~

fiemnur 3.5.16 MC E {P,NP,RP,BPP,PSPACE). Letf be a set o f d i d d d o m a;dddr is eIossd I f b m & t e n u ( A t b ~ ) d ( B , p ~ ) minAver(C,3), t i t e n s ~ i s ( A t B B , j ~ ~ @ p ~ ) -

CKAPTER 3. GENEIlAL TIjlEORY OF AVERAGE CASE COMPJXXlTY 83

ProoE We prove the lemma for the ease C = P. The other cases foRow similarly. Assume that (A, pA) and

(B, pB) are in Aver(P,F). There are two deterministic 'Thring ma&iues MA and MB, and two poIynomiaIs

PA and ps such that L(M,q) = A, L(MB) = 8, Xz.TiiMA (z) is PA on PA-average, and h.TimeM, (z) is

p~ on praverage. We assume without Ioss of generality that both PA and p e are increasing. Consider the

following algorithm N.

begin simulation algorithm N mput z

i f z = X then reject z

hdusu&thateitherz=Ouorz=lu

if z = Ou then simulate MA on input u

else simulate MB on input u

Assume that x = Ou for some tr. It is easy to see that there is an absolute constant c > 0 which does not

depend on the choice of z such that TmeN(z ) 5 c. (lzl + TimeMA (u) + 1). I f we take a d a e n t l y large

constad c, a similar ineqtdty holds for MB if z E lCg.

To complete the proof, we &ow that ~ . T ~ N ( z ) is polynomial on p~ $ psaverage- Flrst set s(n) = c . (n+pa(n)+pB(n) f 1 ) . Notethat sisapoIpmial Foreveryrz 1,

1 L 5 - - f i A ( { ~ l ~ ~ e ~ A ( ~ ) > ~ ~ ( ~ * ~ } ~ ) + ~ ~ b 3 ( ( ~ i T ~ e ~ ~ ( ~ ) > ~ ~ ( l l u l - r ) } )

2 1 1 5 - - &({ti 1 T ~ M ~ (u) > pr(lu1- r ) ) ) + 5 - j b ( {a I I i e ~ , , (4 > P B ( ~ I - r ))) 2

Hence, N nms in po1ynomiaI time on v-average.

3.53 Inclusions and Separations

We shall discuss general separation and collapse restllts among averagecase c o m p l e dasses introduced

in the previous snbsection-

The folIowiug proposition shows a basic sepamtion between averakpcase complexity dasses Dist('D,7)

and Aver(C, *).

Let us recA hm Chapter 2 that a talIy set is a subset of (0)' and b m Section 3.2 that vtpny is the

standard distribution that is positive onlp on (0)': We say a (worst-case) amp- dass C is dosed under

~ u n i m r @ i E , f o t a n p s e t s A a n d B i n C , A 8 B i s d s o i n C .

C&APTER 3. CZMBA.L THEORY OF AVERAGE CASE COMPLEXITY &I

3. Assume that D amtains the s d (0)' and is closed under disjoint u k If REC - D # 0 and

DTIME(O(n)) E C, then Aver(C,F) Dist(D, *).

Proof. Consider a set A in REC - D. Define B = (0)' $ A. Thus, (B, my) M ( D , *) since, othemise, A E D by the closure property of V under $. We now show that (B, vuy) E Aver(C, F) . Let M be a deterministic Thhg machine which computes A. This machine since A is recurs i~ Consider

the following algorithm N for B:

begin simnIation algorithm N for B input z

if z E (0)' then accept

h d y such that z = l y

simuIate M on input y and halt

end.

S i T i m e ~ ( 0 ~ ) I at + c for some absolute constant c > 0, we have

1 & ~ ( ( z 1 T i i e ~ ( z ) > clzl r + c)) 5 &U,.({O~ I Tiime~(0~) > cn + c)) = fiully(0) = 0 < - r

for alI r 1 1. Hence, (B, vay) E Aver@TIME(cn + c), 7) C Aver(C, F). D

As a corollary, we get a result shown by Wang and Belanger [lllj regarding the separation between

Dist(NP, *) and Aver(P, *).

Ccnvuc#ar 3-5-18 [Ill] Aver(P,F) Dist(NP, *) for ony set F of distri6trtions tdi vmy E 3.

Next we shall introduce a notion of weakly C-dexriptiuene~~ for averakpcasearmpkity dasses Aver(C, F).

Deanition 36.19 (Weakly DeScriptivene88) Assume that 'D is a T-time or T-space bounded com-

plexity ciass and C is any (worst-case) c o m p w dass. We d the average compIexity class Aver(D,F)

w d y C-desuiptiae iE, for every problem (D,p) E Aver(D,F), there exist a set C E C and s detr - . tic

oracle Tming machine M such that M with oracle C recognizes D in time 7 on p-awage-

Prpposition 3.5.W Let7 be a set of di&&dhs.

3. GENPLAL m O R Y OF AVERAGE CASE COMPLEXrrY 85

Proof, We shall show only the case (3). Let (D, p) be a distributional prob1m in Aver(BPP,7). Take a probabilistic 'Ihring machine M which computes D ip time polynomial on paverage-

We define two sets C' and GI as follows: Iet Co be the set of strings (0, z, In) such that more than half

of the computation paths of M on mput z terminate m accepting codgurations in 1- than n steps; and

kt Cl be the set of strings (I, z, In) such that more than half the paths of M on z terminate in less than n

steps, It is not hard to see that Co and Cl are PP sets. Set C = Co U CI . Let us consider the f0nowi.g deterministic algdrithm N with oracle C:

begin deterministic oracle Turing machine N with oracle C

mput z

forn=Itooodo

query (0, z, In) to oracle C

if (0, x, In) E C then accept and halt qyery (I,z, I*) to oracle C

if (1, z, ln) 4 C then reject and halt -&for

end.

R e d the d-on of Tiie&(z) in W o n 24. By our a t i o n , the algorithm N with orade C

above repeats the for-Imp Time;(z) + I times before accepting or rejecting x, because Tie&(t) is the

minimal number of steps needed to check if M accepts or rejects z. Note that k.Tirmeb (z) is poIynomiaI on

paverage due to Lemma 33.21. M e m o r e , when the algorithm terminates, its output is always correct.

Eence,

Tiie$(z) 5 d - (Tmeh(z) )+ 1)

for some constant d > 0. Sice kTie;(a(z) is polynomial on paverage, obviously N with C compntes D

m polynomial time on p-- D

Provided that C is a T-time or %space bounded comp1exity class, we say that Aver(C,F) is closed

under weak des- if, for every (C+) E Dist(C,F) and every deterministic oracle ntring machine M

with orade C which is 7-time bounded on paverage, the distriiutiond problem (L(M,C),p) belongs to

Aver(C,F). For exampIe, Aver(P,F) and Aver(PSPACE,F) are d d under weak description-

Lemma 3.531 Fm C € (P,PSPACE}, Aver(C73) is daKd trnda w e d description.

ProoE Here we shaIl show that Aver(PS?ACE, F) is dosed under weak description. Let us first assmne

that (C,p) is in in(PSPACE,F). There exists a deterministic 'Ihcing marhine & computing C using

poIynomial spa^ Suppose S m M ( z ) p(lzD fix an Z, where p is an appropriate increasing po1gnomiaL

Consider any m c oracle Turing machine M which, with o d e C, is poIynomlaEtime h d e d

on p-- We must constrncct a deterministic lhring machine which computes C using space polpnomial

CHAPTER 3- GENERAL THEORY OF AVERAGE CASE COMPLEXITY 86

on paverage without any help 6mm o d e s . The idea is to combine two Turing machines Mo and M:

begin detemhistic dgorithm for N input z

SimnIateMonmptxtz

while sirnulaton do

ifhf queries t then sindate Mo on inpd z

if M reaches a halting confignration then output M(z) and halt end-while

end.

It is obvious that N computes C correctly without any oracles. The tape space used by N on input 2,

SpaceN(x), is dearly bounded by

Apply Lemma 3-3-14 to p ( ~ i e f & ) ) , and we mnchtde that X2SpaceN(z) is polynomial on paverage. This

implies that (L(M, C), p) belongs to Aver(PSPACE, F). 0

Let us now see general separation and collapse d t s among a- comp1exity cIasses. For the propo-

sition below, we &st introduce the supplementary notion of safe sets: given a set 7 of functions, we dl 7 safe if (i) for any ftmctions tf,t2,t3 E 7 and any constant c E N, there exists a fnnction s m 7 snch that

c(tl(z) + t* 0 t3(x) f 1) S(Z) for an x, and (ii) for each frmction t E 7, t(z) > t(y) implies z > y for all z

and y-

ProposiPion 3.5.22 kt C k of thefonn DTIME(T), NTIME(T), RTlME(7), BPTIME(T), ATIME(7),

m DSF'ACE(n for sume safe set 7 of fkndhs on El?.

1. IfC REC, tlrar I) - C # 0 implies Aver(Vt +) - Aver(C, *) # 0.

CHAPTER 3. GElWt4.L TBEORY OF AVERAGE CASE COMPLEXITY 87

a proper cornpIexity core C for D wikh respect to C. Note that C C D. Define the distribution p as follows:

if z E C and 1x1 = n, m = iGp -2-M otherwise

for some adequate normalizing amstaut d Since D € V, we have (D, p) E Dist(D, *) C Aver(D, *).

We next show that (D, p) 4 Aver(C, *). ;Issnme that (D, p) € Aver(C, *) via machine M. Suppose that C is a timebounded c o m p l e dass and assume that Xz.Tiie&) is t on p-average for some t E 7. S i C

is a mmpleJdty core, there exists an integer k > O such that, for all z E C with lzl 2 k, Time&) > t(2[zI3)

because, otherwise, the set (z E C I Time&) < t(21sI3)) is an infinite subset of C. Using this fact, we

bound the probabiity of the event that Timew(%) > t(lx( - r) by

This contradicts the assumption (D,p) E Aver(C, *)-

Note that if C is a space-boanded c o m p I ~ b, use SpaceM(z) instead of TieM(z) in the above

argument.

(2) .Assume Z?' C. Let (D, p) be any distnintional problem m Aver(V, 7). Smce .4ver('D, F) is weakly

Vdescriptive, we can take a set D' m 2) and a deterministic o d e Thing machine M such that M is 7 on p-average and computes D with d e D'.

Notice that D' belongs to C. The dosure p r o m of Awr(C,F) under weak description leads to the

conclusion that (L(M, D')p) is m Aver(C, F). 'This impIies the desired resuIt that (D, p) E Avet(C, T). 0

As a corollary, we obtain the faUowing d t -

Interestingly, the class Aver(NP,3) cannot be characterized by the notion of weak NPdescriptkeness.

Moreover, we are able to show that Aver(NP, *) is indeed diikmst h m Aver(P, *)- The theorem bdow

shows the separation between Aver(P, *) and Aver(Nf, *) (see also [m-

ProoE We shalI eonstraa a d i s t n i decision pmb1en.t (Ap) which belongs to Aver(NP, *) but

CKAPTER 3. GElWRAL THEORY OF AVERAGE CASE COMPLElxITy 88

not to Aver(P, *). F i we choose a nonr-, recmshdy enumerable, tdly set A whose dements are

emmerated as A = {M(so), M(sl), M(sr), . . .) by some deterministic transducer M which always hdts and

outputs some strings For simplicity, we assume withont loss of gen- that if M on input w produces x,

then ThneM(w) 2 1x1 + Iwl. W e also choose the distriiution p defined as

Clearly the distniution p is posithe. We daim that the distributional problem (A, p) belong to Aver(NP, *)

but not to Aver(P, *). We note that if (B, v) E Aver(P, *) and v is positive, then B is recmsitre. Smce A is

not recmsive, it is immediate that (A, p) # Aver(P, *).

We next show that (A, p) E A v e r w , *). Let us consider the nondeterministic l t r ing machine N defined by the foIlowing simple nondeterministic algorithm. Remember that A is a tdly set.

begin nondeterministic algorithm for N

input x

if z 4 (0)' then reject

guess a natural number i ( M y guess y)

simulate M on input si

if z = M(%) then accept else reject

end.

Obviousiy N reeogaizes A. Now fir z and let k = min{i E N I M(si) = x). In the case where x f -4,

TrmeN(x) is set to the shortest length of a q-ecting path by om d-on; thus, it is bounded by a constant

independent of x. Let Q be such a constant. Hence,

Suppose x E A. Wdte D, for the set {i 1 I(%) = z}. On mpnt x, the above algorithm takes at most

steps for some p0Sitif4 constant d which is independent of z. S i i A is a t d y set,

where c is the nomaking constant for p We then have

By Lemma 33.15, we condude that X z . T i N ( ~ ) is p o l p o d on paverage- I3

not dectively caldated and often takes extremely d l probability. NonethelessT most distributions we

encounter in practice do not have auy resemblance to this p. When we discuss practical distributions, we

may d c t our interest to such distributions that give r e h i d y high probability for almost all input strings.

Here we call such distributions supportioc

Dehition 3.5.25 (Swpportive Diiiutiooms) A distriiution p is d e d sappmtke if there exists a

polynomialp such that either ji(x) > rfllzl) or F(x) = 0 for every x. A set of distriutiox~~ is called supporEitte

if all of its distributions are supportive- Let SUPP be the collection of all supportive distributions.

For example, the standard distn'bution u-a is supportive s ine firund(2) > rtive&f - 2-14 > Z - * ~ I - ~

for all z.

For any supportive set FT the average cmmp1exity c h s Aver(NP, 3) is not a large dass. The f01lowing

proposition is a consequence of Lemma 33.1 l(2).

Proof. The r d t fotIows h m the simp1e observation that, for a machine M which is t-time bounded on

paverage, TimeM (z) 5 t(lz(/ji(z)) < t([zl- 2dkl)) if ji(z) > - 2-p(lf 1). D

Regarding the Dist(NP,F) E?Aver(P, *) question, it is enough to focus on supportive distniutions in

F. We introduce an additional terminoolog~r: a set 3 of distriiutious is d e d tome if there exists a positive

and supportive distniutio~t

k n m a 3-5-27 Lef 3 be a tame set of d k b i h h m and assunze that T is &sed under 2-oddition. Then,

Dist(NP, F) E Am(P, *) if wrd only if Dist(NP, SUPP n 3) Aver@, *) .

Proof. (Only if - part) Note that Dist(NP, SlTPP n F) Dist(NP, 7). This inclusion obviody yields

the dab.

(If - part) Let us assmne that Dist(NP, SUPP n F) Aver@, *). Moreover, we assmne that (A, p) is

any distnibutional problem m Dist(NP, F). We shalt show that (A, p) belongs to Aver(P, *).

Smce SUPP n 3 # 0, let us take a supportive distrriution WJ &om 3, and then dehe the default

By the assumption that 3 is closed under 2-addition, this distn'btrtion v k t m t e s a member of F. It is also easy to see that v is supportive, and ccnrseqnentIy, we obtain v E SUPP n F. h m Dist(NP, SUPP n a Aver(P, *), it f o ~ m that (Ap) beiongs to Aver(P, *). Notice that v

p-domhtes p because 2 - P(x) 2 fi(z). Corohy 3-47 helps us condude that (A+) is aIso m Aver(P, *).

In the above lemma, we require 3 to be closed under 2-addition. Later we shall see several tame sets of

distributions which contain the standard distriiution and satisfy this requirement.

3.5.3 Another Characterization

Let us consider another characterization of basic averagecase complexity dasses, Aver@, 3), .4ver(NP,T),

and Aver(BPP,T).

The following definition was proposed by Impagbm [43j. Let f be a function on C and S be a set.

We say that an algorithm A conrputes f with bmign f& if (i) it outputs either an element of rau(fl or "?"

and (ii) on input z, if it outputs z which is not "?", then z = f (z). An aIgorithm A computes S with benign

faults if A computes xs with benign faults. We write TimeA(z,6) to denote the running time of A(z, 6).

Dehition 3.5.28 (Benign Algorithm Scheme) [a] Let p be a distriiution and f be a function on

C'. A polynomial-time benign algorithm scheme for f on input ensemble (wn)nEn is an dgorithm A(=, 6)

such that

(i) there exists a polynomial p such that TirmeA(z, 6) 5 p(!zl,1/4;

(ii) A computes f with benign faults; and

(ii) for all6 (0 < 6 < 1) and aII n E N, if wn is defined, then k n ( { z E Esn I A(z,6) =?I) c 6.

Lemma 3.5.29 [43] Let 7 be a set of diskr'bcrtions+ For a d W 3 h t h d probfem (D, p) toitlr p E 7, the

foUaaing stabmds on CpUioalenL-

Proof. First me shan see that (1) implies (2). Assume that (D, p) is in Aver(P,F). There exisis a

det ' ' "c TEng machine M such that L(M) = D and time^ is poIynornia1 on paverage. By Lemma

338, there is a polynomia p such that k n ( { z € Xsn I T.une~(z) > p([xi - r))) < $ for all n E N and all

r > 0. We define the desirrd agOrithm A(z,6) as fall- on input (z,6), A &dates M on input z for

p(Iz1/6) steps and outputs &Tb if M faiIS to halt- Tbis aIgorithm is a benigu algorithm for D on (en)nEn since

To see that (2) implies (I), we assume that A(z,6) is a benign aIgorithm for D on inpd ensemble

k ) r n . We also assmne that A nms in time p(lzl,1/6) far some fixed polynomiaI p. We may take a

C&QPTER 3. GENERAL TBEORY OF AVERAGE CASE COMPLEXITY 91

may polynomial q instead of p, because we can assume that p is maeasing (LC, h.p(n, 6) and Xt.p(m, z)

are incfeasing for each fixed 6 and m); then setting q(z) = p(z, z), we have p(lzl,l/b) 5 q(lzl/b). Moreover,

we specify the form of q(z) as z*C + d by taking sut€iaently large k and d We then asmme that the benign

dgorithm scheme A(x, 15) runs in time (1z1/6)' + d

Now let us d d e r the following algorithm M that computes D:

begin detr ' ' ' 'c aIgorithm M for D

inpat2 for i = 1 to oo

simulate A(z, l/i)

if A(=, l/i) outputs "?" then go to (*)

output A(x7 l / i) and halt

(*I end-for end.

This algorithm M a d y computes D. Let us define s(t) = + dz for all z. For any string z of length

5 n, if the algorithm M hats within the first r iterations of the for-loop, then

In other words, if T i i e ~ ( z ) > s(Iz1- r), then A(z, I/r) =?. Hence, we have

By CorolIary 333, h-Tim&) is polynomial on paverage. 0

We observe that the qnantik charaaerizatiori of nond ' ' ' tic and probabilistic Thing rmchha

hdds alsointhea-setting. E&caIl that,forinstance, allsets inNP can b e d m a d m i d i n terms

of an existential quant5er and sets m P as fo1Iows: a set A is m NP if and onlp if there exist a polynomial

p and a set B E P such that A = ( x f 3yUyl Ip(1xl) A (z, y) E B]) [118]. l~ the fbIIowing, we shall give a logid characterization of the dass Avet(NP, 7).

CHAPTER 3. GENEEbU THEORY OF AVERAGE CASE COMPLEXfm 92

Proof. The proof is straig6dorward and follows from the standard techniqne of encoding nondeterministic

computation paths into strings and from the fact tbat I\z.Time~(x) is polynomial on paverage. 0

Fm we need an amplification lemma

Lenuna 3.5.51 (Amplification Lemma) Let p be wry diskibtrtion, A a set, and d u positive finac-

tion con@&& in polynomial time on p-olrerrrge. Assumc fhot u rrmdomized %kg machine M sdisjies

RM[M(z) = A(z)] 2 ? + & fm all strings z. Then, thae ezists oncrtha rondomiad machine N

suchthat

where p x u u y h t e s the distn'hdim pf de&ed by $(z,y) = b(x) fiuy(y) for On p i m (2, y).

Proof. La us assume that ks.TiieM(z; s) is t on p-average for some increasing polynomial t. -4ssume

also that d is computable in time polynomial on paverage For simplicitp, we assume that d(z) 2 5 for all

z. We set h(z, y) = max{l,lzl+ lyl) and let p(z, y) = 2 d ( ~ ) ~ h ( z , y) for all pairs (x, y). Fmt we show that

p is polynomial on p x uay-average- For simplicitp, we write p' ffot p x p a J r in the following argument.

Proof of Claim Remember that d is polynomial on gaverage. Suppose that d is q on p-average for some

poIynomiaI q, and dehe q' as q'(r) = 2 q ( ~ ) ~ z . Obviously q' is a polynomial. Fa r a r b i i y such that

1. r 1. ~a E, = { ( z , ~ ) I > q(r - I~I>).

Hence, p is polynomial on $-average I

Smce p is poIynomiaI on $-a- m can take a poIynomid q snth that p is q on $-average- The

desired randomized machine N is de6ned by the foflowfng aIgmithm:

begin randomized aIgorithm for N

inpat ( ~ 2 ~ 1 if y fZ (1)' then reject

CKAWER 3. GEiVElUL TBEORY OF AVERAGE CASE C O W -

let A c c s O and& :=0

for i = t to p(x, y) do

simnIate M on input z

if M halts in an accepting state

then Iet Acc z= k + 1 else let Rej s Rej + 1

end for

if Acc 2 Rej then accept else reject

end.

Notice that the random-input domain for N is

The nmning time of the machine N on input (x, y ) with random input r, T i e ~ ( z , y; s), is at most

for some absolute constant c > 0.

We first show that the random ftmction ~ z y s . eiv' T i i e ~ ( q si) is po1ynomiaI on pt-average To &ow

this, we set t&) = (q(z) - i) - t (z) for each number i E N, and we also set t'(z) = c - q(z) - t ( z ) + 1. Notice

that t'(z) =to(z). Let r > 1 be fmd. k D+= {(z,y,s) Err I y E {l)'Ap(z,y) Iq(r-h(z,y))). By om

as~nmption, it is obviow that fit(=) < &. h is enough to S ~ O W that fiiN (((2, y, S ) E Dr I T i m e ~ ( ~ , y ; S) > t"(r (1x1 + lgl))}) < l / r for some polynomial t".

Let x be b e d and set n = 1x1. For this z , write P,$ for P r , m e ~ ( z ; s) > t(L - n) I s E r ~ ( z ) ] . .Ah

let y be lm. Set f(u, w) = c (t(u) - q(w) + 1). Oboiottsty t' is a polynomial.

hf of CIaim The estimation is d e d out as fonoars:

CHAPTER 3. GENERAL THEORY OF AVERAGE CASE COMPLEXITY

Repeating this estimation, we reach the oondnsion that

The Iast inequality comes from the .%?sumption that p(z, y) < q[r(n + m)).

We now define t"(z) = c - (t(@ - q(z)) - q(r) + 1) for all z. Notice tbat

t"(r(n + m)) 1 t'(4n3m2r - q(r(n + m)), r(n + m)) > t'(4n3r - q(r(n + m)), r(n + m)).

Hence, the rest of the dcuIation is carried out as f o h . For any real number r 2 1,

Therefore, ~ S . T I r n e ~ ( 2 , Y;S) k p0lpmid On p X I&@-

Next we examine the error probability p, that N on inpd z outputs a wrong answer; that is, RN[N(z, lm) # A(z)]. Wiite r, for R M [ M ( x ) # A(z)I, for ampatison, The error probability p, does not exceed the pmb

abi& that M on input z does not compute A(z) correctly m more than (= d ( ~ ) ~ h ( z , y)) independdeut

trials. Hence,

CHAPTER 3. GENER4.L THEORY OF A'CrERAGE CASE COMPl2XR"Y

since (1 - $)n e-4 and 4n2 5 en for any number n greater than 4. Therefore, we obtain Pr,[N(z, lm; s) = A(2) I S E rN(Z, lm)] 2 1 - 2-lrl-m- 0

Cmllarg 3.5.52 Let p be a dktddion and let d be a non-ren, valued jbction computable in polynomial

time on p-merage. Asswnc that, for a given set A, a rmzdOmited Turing machine M satisjies PrM[M(z) = A(x)] 1 $ + & forullstrfngsz. Forewryfundiont w ~ i 9 ~ k i n p o l y n o n r i a l t i t n e o n p -

attemge, there ezists a randomixed %ing rnacirine N which runs in polyrmomral time on p-memge such that

Rrv[N(z) = A(x)l 2 1 - 2-f(r) fm all sdrings x.

Proof. Amme that M &es RM[M(2) = .4(x)] 2 f + & for all x. AppIy the AmpBcation

Lemma, and we construct a randomid ' h h g marhine N such that Xzs.Tiie~(t, y, s) is polynomial on

p x u~Y-average and Pr,[N(z, lm; s) = A(z)] 2 1 - 2'[zl-m.

Given a function t computable m polynomial time on paverage, we define another randomid M g

machine N' that takes input z and then simulates N on input (x, lt(')). By the definition of N', the success

probability RN[N(t) = A(z)] is at least 1 - 2-w-t(z) > - 1 - 2';(=).

To compkte the proot, it dices to show that ks.Tiie~p (5 s) is polynomia on p-average- Notice that

for all z. As a result, by Lanma 3322, h . T i e ~ r (z, s) is polynomiai on paverage- 0

Pmpodkn 3.5.33 Let 7 be a set of &dtdhs . For every set D and every d i d d u t h p in 7, the

fouorping statements me eql&aIatt.-

1- ( 0 , ~ ) ~Ave.r(Bpp,F)-

CZAElER 3. GENERAL T&EORY OF AVERAGE CASE COMPLEXITY

(iii) Mo and k s M 1 ((x, h(x)); s) mc potymdaf-time boundkd on p-menzgc; mul

Root By applying C o m ~ 3.532 to (I), we immediately obtain (2). To see the implication h m (2)

to (3), set q = h.2 and take M gtisfying conditions (i)-(ii) in (2). Let us dehe Mo as the randomized

Thing machine that works as foIlows: on input z, it outputs 0 without flipping any coins. Moreover, let MI works as foIIm: on input (z, p), simulate M on mput 2. Let d(z) = 4 and h(z) = A (the empty string). It

easy to check conditions (i)-(iv) in (3).

Now it remains only to show that (3) implies (1). Assome that (3) holds. There are n r ~ , randomized

Thing machines (MOT MI) satisfying conditions (i)-(iv) in (3). By the .4rnplification Lemma, we can modify

MI to another randomized l h h g machine Ma,, so that R, [Ma&, 1 &, s'; s) = A(=, s')] 1 1 - 2-*, where

A is the set accepted by Mt ((2, M o ( ~ d))).

b e e round robin algorithm for N

input x (say, n = 1x1) let s = (empty string)

f o r k = l t o c u d o

for j= l to kdo

caU subroutine CHECK(z, s, j, k)

end for

end for

end.

Here is the subroutine CHECK(%, s, j, k):

begin subroutine CHECK(z, s, j, k)

for aII s' such that Is'[ = j if either Mo(z; s) does not halt or it halts

without using up all random bits m s' then go to (+)

(Assume that Mo(q J) hats after asing up all random bits m $1 Iet y be the output of Mo(z; s) compute e F 2j + 2

sindate the kth step of M, on mput (y, le) using random seed s

(If this step consists d a coin-@ping state, then Bip a coin and let s = sb,

where b is the outcome (0 or 1) of the coin toss.)

if M,, d e s a halting state then output M-&, y), le; s) and hait

(*I end for retrmt

Notice that random seed s is shared commoaly by an computatiiom of M' in the subroutine CHECK-

CHAPTER 3. GENERAL THEORY OF AVERAGE CASE COMPLEXlTY 97

We shall perform worst-case analysis on the running time of the above algorithm, On inpnt z dong with

random seed r, N takes Steps at most

where f is an initial segment of r. Notice that the bct ion &.21h(=)1 is polynomial on paverage by Lemma

33.6. Hence, k-TilllleN(~ r) turns out to be po1ynomial on paverage.

Next we shdl show that R N [ N ( 2 ) = A(%)] 2 t c For each random seed, the success probability does

not dedine below 2/3. Thus, the probability that N(x) coincides with A(z) is at least 2/3- Therefore, (D+)

belongs to Aver(BPP, 7). o

We give a remark to tbe above propositi01~' if MI makes one-sided errors (i.e., ( D , p ) E Aver(RP, *)I, then we can replace the term $ + &J by &.

Further Topics

This section cibases sewral topics which we have not covered in the previous sections.

Rmmiag Time of Nondeterministic Machine. Let us considr the running time of nondeterministic

M g machines- Wang and Belanger [Ill], Schnfa and Yamakami [97], and Karg and SchuIer [48] disepssed

a dose connection between the P =?NP qyestion and the Awr(P,F) =?Aver(NP,F) qnestion based on

the model of d d e d n o n d e t d c 'Ihring machines. As we have seen, this model guarantees that aII

compdation paths in a computation tree of the clocked machines are of the same length.

Here we consider the model for which the nmning time of a nondeterministic flrring machine is d&ed

to be the length of the shortest accepting computation paths when it accepts an inpnt, or else the length

of the longest rejecting cornpubtion paths. We call this restpiction the strid nnnring t h e cnterunr . - for the

machine

To avoid c o ~ o n , we use the notation Aver(NI", 7) to denote the awrage-case cornpleexi~r cIass defined

m terms of nondeterministic 'k ing xnachines, with the strict rrmning time criteria, which nm m p01ynomial

time on average-

For this new dass A p e r W , F ) , we are able to p m that Aver@NF,F) is weakly NPdescriptive, and

thus, P=NPimpIiesAver(P,F) =Aver(NP',F) for anyset Fofdistriiutions-

Aaof of C&im Let (D,p) be a distriiutional pmb1ezn in Aver(NP', F). Let us consider a noxtd-c

' k ing macbe 1\6 which computes D ia polynomial time on pawrage.

Let 6 be the set of sttifl&9 of the form (Op, In) such that M accepts x in Iess than n steps, and Iet Cr bethesetofstringsoftheform (17z7P) s n & t h a t t h e r e & m p n t a t i o n ~ d M o n m p d x w h i c h ~

CHAPTER 3. GENERtQI; THEORY OF AVEZUGE CASE COMPLEXfm 98

Ionger thaa or equaI to n. Now write C for Co U CI. Let us consider the fonowing deterministic algorithm

N with orde C.

begin deterministic oracle Turing marhine N with o d e C input z

forn=l tom do

query (0, z, In) to d e C

if (O,z, 1") E C then accept and halt

query (1, x, In) to oracle C if (1, x , 1") 6 C then reject and halt

end-for

ertd.

On any input z, this algorithm N with oracle C goes through the for-loop TimeM(z) + l times until it

terminates, Notice that when the aIgorithm terminates, its output is always correct. Hence,

for some constant d > 0. Smce k.TieM(Z) is polynod on paverage, obviously N with C computes D in time poIynomial on p-average- I

Heuristic Polynomial T i e . ZmpagliazLo [43] introduced a new notion of "hetnistic plynod-timen

class. We rephrase his d&tion in our setting. We dehe the heuristic polynomial-time class Heur(P,F)

as foRowsOWS

An dgoritlrnt scherne for (D,p) is an algorithm A(z,d) such that k,({z E Csn I A(z,6) # D ( z ) ) ) 5 6 for aII 6 > 0 and a11 n E N We say that a distribntid decision problem (D,p) is in Heur(P,F) if there

exists a polynolniaI-time deterministic algorithm scheme for (D, p).

Obviously Aver(P,F) E Heur(P,F). It is an open probIem whether Dist(NP,F) C Heur(P,F).

Chapter 4

Feasible Distributions

4.1 Introduction

A- d y s i s m g e n d is sensitive to the choice of distniutions, so that we need a careful study of

the behaviors and properties of individual distriiutions. For example, to construct a better dgorithrn which

nms f& on merrrge, instances which occur with high probability &odd be s01ved QtrickIy.

In statistics, Poisson distri'butions, for example, are commonly used for the analysis of events. This type

of distriiution is apprmdmab1e by some appropriate po1ynoW-time algorithms.

In this chapter, we shall f- on fecutile d i & d & h s which are either YeasliIy comptttabIen or "fksi'bbly

pmiua%Ie? This chapter wiII introduce two Merent categories of feaslile ciistri'butioris: pofgn-d-he

wmputubk dC&iMiow and pofymmial-time smtplble &dt&w. The former category was considered

by Levin [g0l and fonnuked by Guretrich [36j; the Iatter was introduced by Ben-David, Chor, Goldreich,

and Luby [9] and has been studied in ayptqraphy.

FoIIOtKing KO and Eti-'s [55] de6nition of p ~ l ~ t i n r e cmptable d fmctionr, Garevich [36j

took a shdar step toward distn'btrtim. Be d e d a distribution "cornpntabIe in po1ynomiaI-timen if there

is a poIynoxniat-time approximation aIgorithm whose outputs asymptotically approach the values of a distri-

bution within an exponentiany matt factor. Section 42 wiIl discuss such diskiitttio~~ which are computable

(or more acamtdy "apprOOdmablen) in pcdynomia time by deterministic %ring machius-

In practice, rather than spedjkgfuu distri'btxtion~~ researchers often loosely deke distrritttions by sim-

ply specifping associated serm-drstn - - %utions. We note that aII polynomiaEtime computabIe semi-distrrr'btttions

are e k t i d y enumerated, whiIe associated fidl distriiutions are not- Remember that our theoq is M

on jkl2 distriiutions, and therefore, whenever we use semidistributions for the purpose of defining /Glf distn'butions, we must puantee the existence of snch j U distri'butions that are proportional to the semi-

distriiutions almcst This process is called a - - ofSemi-distninti~- U d o ~ , not all compntable semidjstributiom axe normaGd to fid distnitrtions of the same c o m p 1 e . Section

4.3 d s h o w t h i s negativemi&.

CHAPTER 4. FEASIB~DISTRIBUTIONS 100

As Gurevich [36j pointed out, the reader may be cantious of the fact that the P-computability of a

density function does not imply that of its asmiatd distriiution de s s P = NP.

One of the most important results in Seaion 4!4? is the Distribution Controlling Lemma proven by

Gafevich [36] and by Belanger and Wang [6]. This lemma enables us in Chapter 5 to prove the existence of

complete distniutiond decision problems for Dist(NP,P-comp).

We shall turn our interest to instances which occur with low probability under most distriiutions com-

putable in polynomial time- These instances are called rare skin@- A set S is d e d polynomial &rare if,

for any polynomial-time computable distniation p, the set { x E S [ i ( x ) > 2-c(l*I)) is finite.

Another category of feasible distributions is amplabk & d t d h s which were first introduced by Ben-

David, Chor, GoIdreich, and Luby [9] in their 1990 confkrence paper. Samplabiity, which is often found m

st'atistical physics, is esrentially a form of psendc+randomness. For example, Ben-David d d [9] showed that

if psendwandom generatom exist, then polynomiaLtime samplable distributions are hard to approximate

in polynomial time. This result wi l l be extended in Section 4.7. In Section 4.4, sampIing algorithms and

samplable distributions wiI l be defined.

Major Contrriutions. Most of the material in this chapter comes !?om Yamalolmi [119].

The formulation of P-samplable distrriutious given in this chapter is merent from what has been defined

by Ben-David, Chor, Goldreich, and Luby [9] and by Schder and Watanabe [96j, and therefore a11 the proofs

are altered.

Given any P - c o m p d l e distrihtion p, Lemma 4.2.8 presents a a a e n t condition of a function f so

that the composition el becomes P a m p a b b I e

Theorem 42.14 shows the existence of poIynonriaIIy C-rare sets of distriiutions. This theorem dates to

Kolmogorov compIeJdty. The theorem actuatly proves that, for any hcmsing, unbounded function s, if l

satisfies 5 logn 5 C(n) n, then the set whase dements are uot in KT[S(n), p(n)+g(n).'wnJ is po1ynomian:y

C-rare

Proposition 43.1 shows that there exists a P - c o m p d e Semi-distn'bution which cannot be normahzed

by a standard method to a fJI distniution computable in poIynomiaI time- In addition, CoroIlary 43.3

gives a &dent condition for P-compntabIe semi-distniutiom to be normalid by a standard method to

distn'butions of the same complexity.

Proposition 42.12 shows that there ex& a positive PcomputabIe distriiution p and an inQeaSing,

exphonest ftmction f m FP such that prl is not pdomioated by any Pamputable distxiiution.

In Proposition 447, IP-samp is shown to be included in P-samp.

Theorem 4.413 asserts that ~ - - s a m ~ peqd P-samp- This redt is a counterpart of the d t

BPF = B P P in worst-case c o m p w theory.

Oneoft6emostimpaaanttheorem~inthis~terisTbeoremrL5.~ whichshowsthat P=PP ifand

only if P-comp = P-samp- Not aU P-samplabIe disbibutiom are therefore computable in polynomial time

d e s P = PP.

CHAPTER 4. FEASIBLE DLSTRIBUTIONS 101

Theorem 4.7.3 shows that if every P-samplabIe distribution is pdominated by some P-compntabte dis-

tribution, then every NP set is nearly-RP.

CbmIIary 4.7.7 shows #P-comp peguals P-samp if NP E BPP. This remlt foffows Proposition 4.7.6,

a careful rn6cation of a result proven by Schuler and Watanabe [96].

Proposition 4-73 shows that IP1-samp is not avp-included in P-comp unless P = RP. Moreaver,

Theorem 4.7.12 shows that P-samp is not pincluded in P-comp ndess P = NP.

4.2 Computable Distributions

It seems natural to consider the sets of all feasibly "computablen distriiutions. Gnrevich [36] first adapted the

idea of the P-amputable red mnnbers used by KO and Ftiedman [55] to define polynomial-time computabii

of distributions. This section win farther explore distributions comprhble in polynoinial time.

4.2.1 De5ition of Computable Distriibutions

Any distriiution treated in this thesis is actually a real-valued function with C' as its domain. Gurevich

has applied the notion of ~~~tinu computcbility of real-valued functions (introduced in Section 2-7)

to distriiutions Here we shall give a more general notion of t-time conrputability and t-space compprtabilidy

of semi-distributions for a fnnction t, which is based on Gurevich's approodmation scheme.

Let us recall how a deterministic Thing marrhine M equipped with an output tape computes a real

nmnber: the machine achtatIy outputs a binary string w, which is interpreted as the dyadic ratio& number

2 1 wi 2-', where w = w ~ q w1,,,l and each Wj is a bit (in (0,l)). We say that M opptoninates

red-value p(z) if Ip(z) - M (z, 0') 1 1 2" for all natmaI numbers i.

Deflnitioa 4.2.1 (Computable Distr'butioas) [55, 361 Let t be a function on N and 7 be a set of

frmctiom on N.

1- A semi-distriiution p is recuffitre (or cornpukr6te) if there exists a detetministic Tming machine M eqnipped with two input tapes and one output tape, which computes namdy, on input (z,oi),

I&)-M(z,oi)l 52-'for ollstringsz~C'and d l n u r n b e r s i ~ ~

2. A semidistriiution p is t-time wmputable (t-spuce computable, resp.) if there exists a deterministi c

'Ihring machine equipped with two input tapes and one output tape which, on input (z, Oi), computes

~ ( 4 time (rising space, tesp-1 w, i)- 3. A semi-distniution is 7-time oomputable (7-space amrpcdc6lq resp.) if it is t-time (t-space, resp.)

comptitabk for some t E 7.

Note that, by Lemma A.4, the Turing m d h e M in -on 4.2.1 uniquely determines the disttiitttion

P-

Figme 4.1 itlrdrates an asymptotic ap@ of the value M(X, OL) ta the vdue of a distribution.

h x

Figme 4.1: A distribution and its approximation

The reader should note that if we relax our dehition of computabii by requesting only the condition

that lp(z) - M(z , o')( I zi for "almost aII" z and i , then semidistriiutions are always normalized to fiJI distriiutions of the same complexity- Here we mean by "normahtion" the existence of the fiJldistrriution

pr satisfying that fir(%) = i(z) fm almost all string z. (Another way to normalize a semi-distriiution to a

distniution is to multiply by a constant c so that c p(x) converges to 1.) This subject wi I l be discussed

again in Section 43-

Note that, for each x, p(z) is a t-time computabk TeaI number. Thus, if p is t-time compntable, then

{ p ( ~ ) ) , ~ p is an inQ.easing sequence of t-time computable real numbers which c o n q to I. &a the

conwse may not be true in generat

We next introdu(3e sets of distniations which are M y computabIe.

MOP 43.2 (Computable Distriiutions)

1- The notation RECcomp denotes the set of ail recnrsive distributions.

2- A semi-distnintion p is c a M pdynomiol-time eonput46le (P-computable, for short) if t h e exists

a po1pomiai p such that p is ptime computable The notation P-comp denotes the coIIection of aIl

P-computabIe distnintions

distributions which are EcomputabIe, EXP-compatabIe, and L-computable

A simple exampIe is the standard distniution v w . It is easy to see fiom its definition that u-d is

L a m putable.

We can aIwaps assmne that M on input (z,Oi) outputs at most i + 1 bits. To see this, let M' be the

machine that m r b as foIIm: it trims the b i i fraction of M(z,Oi+') by rounding it down to i + 1

bits; if the last bii is 1, then it adds 2"'l to make the number of bits at most i + 1. This implies

IM(z, 0') - M1(z, o')( 5 2"", and thns, we get

Example 4.2.3 The nr i fm d i d r h t h over a finite set of integers is P-computable, where we iden*

integers wi th strings. For two distinct natural nmnbers a and b, we define p[,,y as follows:

ExampIe 43.4 Another common dkrete distribution is the P&on distribidion with parameter X > 0

that is defined on (0)' by

for every number n E N We here consider the ose that X is a positive rational nmnber. Notice that

the kth approldmation of the d u e 5 can be computed by a Turing machine m time polynomial in n

and k. S i l y , by considering the power series for e'A, the number e-A is Pamputable. Hence, p is

P-computable

Ben-David, Chor, GoMreich, and Luby [9], on the other hand, define polynomial-time c o m p ~ d

a distribution p m such a way that, for aIT z, the d u e of p(x) is exactly expressed m binary by some

polpomia-time bounded transducer- To distingnish this concept from ours, let us call snch distninti011~

stridly P-cmputde and denote by strict-P-amp the dass of all strictly P-computable distniutions. By

this definition, if p is strictly P-computable, then each d u e p(z) is either 0 or at least 2*lzl) for some

polynomiai p; thns, p is supportme- This Implies that the set strict-Pamp is supportive. SrmiIarly, we

use the notation strict-IiECcomp to denote the set of all distributions which are stridly amptable by

d d e r m i n i s t i c ~ g ~ p l 9 .

It is worth temarking that if a distn'buth is 7-time computable, then the densitp hct ion is aIso 7-time

computable. Theccmverse,hoareper,maymtd~haldsinceitisshmbelooPthatifP#NP,then

there is a P-amnpntable density fimctkm whose associated distribution cannot be computed m p o l y n d

time This d t is dne to Gnrevich [3q.

Proof. Asnrme that P # NP. Take a set A m NP - P such that A = (z I 3y[ly[ = 1x1 A (z, y) E Bl) for

some set B E P. Let r(z) = 0 i£ 121 is odd; otherrioc, @(z) = *i, where lzl = 2n. Clearly v is in P-comp. Now

define the desired distribution p as foffows:

The following summation shows that this p is truly a distniution:

Note that, by clehition, z E A if and only 8 p(zlOlfl) - p(d~*~+') # 0. If p E P-comp, then we have

A E P, a contradiction Therefore, p is not in P-comp. 0

Another observation given by Gnrevich [36j is:

(a) kM(z,Oi) is increasing (ia, z < y implies M(x, 0') 5 M(y,Oi)) for aery i E N.

Proof. Let p be a distniution m P-comp. By definition, there is a polynomial-time Turing machine N such that Ip(z) - N(z , O')ls r'. We defhe the deterministic lbbtg &e M as follows:

else if N(c, O)) > A(€) then set A(c) F A(b)

else set A(c) .r N(c, ok) i f z ~ c t h e n s e t b ~ c e l s e s e t a ~ c

end-repeat

end.

To verify that M satisfies the required conditions is easy- 0

Our apptopdmation scheme is obviously d i f f i t b m the Pwting point model, another popular approxi-

mation scheme given by a pracisia poafirrs-poinf reprw&ah+ becaase, even if p is P-computable, the kth

bit of p(x) may not be computable in time polynomial in 1x1 and k. Nonetheless, the numbers represented

in a finite precision model does not satisfy the law of associativity. AIthough we cannot know any exact pre

cision of the value p(z) in polynomial time in general, it is &dent to know an apprcadmate value M ( z , 0')

within an exponential factor of its true value p(z) when we consider the average behavior of algorithms

under the distniution p

Part of the following usefnl observation was made by Gurevich [XI and by Wang and BeIanger [112].

This lemma is the basis of the proofs of average NP-completeness shown in Chapter 5.

temmo 4.27 (Diiiution Controlling Lemma) [36,112] Let p be a distribution in P a m p .

I . There ezists a positive djddtdion v in strict-Pamp 4 that, for may z, the d u e t(z) hus ut

most 2121 + 4 binmy digits, and 4 - -(z) > F(z) ha.

R o o E (1) Asnrme p E P-comp. I€ there is an z mch that p(x) = 1, then, since the set (y I y 5 x) is hite,wecaneasilyd&evwhich~esthecbim. Intherestofthepmof,wethusassnme that p(z) c 1

for all z.

There exists a polynomial-time lhring machine M such that - M(z, Oi)I < 2-' for aU z and i. For

each string x, take i = 2 1 4 + 1 and let N(z) = ~(z,02l~+I+')- CleatIy we have lp(z) - N(z)j c 2-21f+l-1.

We trim the binary output of N(z] by rounding it down to 2(2+( + 1 di& if the last bit is 0, then cross it

ont; otherwise, add 2-21z+l-' and then cmss out the last bit in order to make the number of b i i at most

212+1+ 1. Now let N'(z) be the result obtained by this process. If z = A, then Nr(z) outputs the &st two

bits of the b i i &don of N(z). Note that [N(z) - Nt(z)l 5 2 * I z + l . By the definition of N', we have

for alI x # A. Hence, b(A) - N'(X)[ < 1 and b(z) - Nr(z}[ < 2-21z+l for all nonempty strings z. Note that Nt(z) i s q r e s e i m binarpwithatmost21z*J + 2 ~ 2 ~ x ~ + 4 ~ s i n c e lz+ls1z1+1+

To get the desired distribution, we dehe

+(N'(A) + 1) ifz=A, P ( 2 ) =

) (Nr(z) - N'(z-) + 2-21z1C1) otherwise

Note that ti(%) > 0 for aB z since N t ( z - ) has at most 2121 bits h m our definition, it is easy to obeain

that &(z) = 1 + N'(z) + 2 &,, 2-4.1 for aII nonempty strings z. Hence, -

For the empty string A, &(A) = N' (A) + I > ji(X) - 1 + 1 = fi(X). For the other strings z,

To show that 2 2-V(+)I, let b(z) be the position of the leftmost digit 1 in the binary fraction of q. That is, Tb(r) 5 < 2 -2-b(t). We daim that 6(2) 5 d(=) +I. Assmne othexwise. S i N ( z ) 2 2-d(z) .

and 9 < e ( z ) C l , b w N(=) - > *(+I -fl(t)+l > - z--d(t)-l . Then, N(=) - 9 > - 2-~(I4)4,

a c o d c t i o n . Using the fyt that ti'(=) d(z) + I, we maciude that 9 2 2d(r) 2 2+(+)-l > - 2-lf(=)l.

0

CHAPTER 4 FEASIBLJ3DISTRIBUTIONS 107

Proof; We assmne that p is a P-comyutable distriiution witnessed by a deterministic polynd-t ime

W g machine hf. Assme a h that a function f is P-computable-

Let us dehe another function g as g(z) = max{z I f (2) 5 z) for each z.

Proof of Claim. Take an appropriate poIynomiaI such that If (z) 1 5 p(lz1) for an strings z. Using this upper

bound of f, we compute g as foIlows:

begin deterministic algorithm for g

mput z

compute the minimal n such that 121 5 p(n)

let a = X and let 6 7 1"

=Peat (binary s'xuch part)

i f a = b then output a

take the mhimaI c such that Il{z I a < z c c)ll > Il{z I c < z 5 b)ll

iff (c) 5 z then set a c else set b := c

end-repeat

d.

The b i i search part of the above dgorithm takes at most O(n) steps, and as a r d t , the algorithm needs

polynomialIy-many steps to compnte g(z). I

Write v for the distriiution k.p(g(z)). We then have:

Let us d&e the deterministic Turing mar)linp N as folI0ws: on input (z, @), simnIate M on input (g(z), @)-

Simce g is Pampatable, the machine N is a h PamputabIe To compIete the proof, we must check the

a~proodmabi~ i tpof~ . This i s~eena~f~now~:

In aqdexity theory, a diaganaIizati011 argmnent is one of the most powerfd and popular tools for

disc.nssinrd a separation of two cornpIexity Note that such an argument is based on the existence of

an e k t b e Cte, recnrsi9e) emmemth of dall elements m question. We already know that, for exampIe, it

is possibIetoewmeateaIIP+mmputabIesetsmane&ctbeway.

The reader may wonder if one can effectively enumerate all distributions. S i P-computable distri-

butions are charaderized by %ring machines, we must emmerate Thing machines which compute such

distributions. It seem, however, difEicult to coMtrnct such an enumeration, because there are two objec-

titles: (I) we must cbedc whether the chosen M g machine, say M, guarautees the convergence of its d u e

to 1 (i.e, lim,+= M(x, 0') = 1) (2) we must dectively check whether the machine satisfies the convergence

scheme, IM(z, 0') - M ( z , Oi)l 5 2-' + 2-j.

To avoid checking the con-ce scheme, let us turn our attention to strictly P-computable distri-

butions. By Lemma 4.2.7(1), we can bound an arbitrary P-computable distribution by mother strictly

P-computable distribution with a constant fsctor. For most applications, we thedore consider strictly P- computable distributions. Can we now enumerate all strictly P-computable distriiutions ? We have to

resolve the problem of checking the convergence of a machine. A d y we do not need to restrict attention

to foll distributions. Schuler [92] kt pointed out the existence of an effective enumeration of all strictly

P-computable semi-dktributions.

Proof. The method of proving this lemma is basidly the same as in Lemma 42.6. Tdce any effective

enumeration of polynomial-time detaministic ' b h g nxxhks, say (M*)iEx. We modify each machine

Mi into another machine Mi which is increasing by an algorithm simiIar to that presented in the proof of

Lemma 42.6 (by replacing N by Mi and M by Mi)). The seqnence {Mil iEN obtained by this modification

as0 becomes an dixtive enumeration of deterministic M g machines, and each M: computes some semi-

diEtrriution m polynomial time. U

Ftom the above theorem, we remark that we cannot exclude the tritriat semi-distriiution kom the en*

mefation.

4.2.2 Rare Strings and Rare Sets

This section will consider input strings which occur rarely for most P-comptrtable distnitrti~ns~ These strings need special anention beguse the set of such strings consists of the most rncult instances for most

aIgorithms to work on in polynomial time

De5ition 4.210 (Ban St-) Let F be an enumeration of semi-distniutions, say F = {h, w , . . .}, and let k,s be fanctions k n N to P. A string z is rmr mah rrspcd to ( k , ~ , a if h(z) < ~ - ~ ( t 4 ) holds

for all i < k([xl). Let RARE(k, s, 7) be the coflection d dl rare strings with respect to (k7 s, 7).

Note that the rareness of a string depends on the etnrmeratian F.

Proof. We 6rst show that, for any integer no > 0, there exists an n with no .I n 5 3% + 6K(no) such .yW that, for each i < k(no), 11{z E IF I B(z) > 2-8(n))n < %. Asrmme o&emise. Let r(n) = 3n +6k(n)

and define 4 = {z E Cn ( h(x) > 2-'(")). Take no such that, for all n with no n 5 r(n& there exists r no)--m+l an i < k(no) satisfying II&I( > k(n) - Pcn)/n. Hence. at least many n's the condition

11411 > k(n) - 2.(") fn for some j < k(w). Let c = Ik~")-')Mf'o)+')+nO k(no

. S i r(n0) > -,, we have

(3) 2 L Then,

The lemma immediately folIows fiom the foUowing meqnalitp:

We can construct a P-computabIe distribution p and an exphonest, P-computab1e function f such that

no distriiutions in P a m p dominate ~ c f l - Recan h m Chapter 2 that p p-&mh&s v if and only if

there is a pbounded function p h m C to P such that p(z) -ji(z) 1 t(z) for an z.

ProposiPion 4.2.f 2 [I191 Them aists a pusilioe P-amptable d i d d t d h p and an inacruing, ap

honest function f in FP judr that ~ l p is not p domkradedbyanyP~tediskiburions.

Proof. We first dehe q as folI0ws:

Let b(z) = $5-(z) + %(z). The distriintion C( is positive and obviously P-compntable For every n 2 2

and for z =v6, we have

y such that logn 5 lyl I 9logn and jyIk-' - &(v) < P(On) for all i < Iogn and all integers k with

15k5 w&&n,wheren=min{r~f15 PI <(r+l)6}3n>213;otharise,l& f(r) =I.

b c t i o n f is ddefined- To see this, consider the case [zl= n6 hr some n 2 213. By choosing logn

as k(n) and n + logn - 2 Ioglogn - 4 as s(n) ir: Lemma 4-2-11, we know that there exists at least one rare

rtMg tO (k,s,F) with 4 n 5 ]y{ 5 9 lop , i.e, fii(y) 5 2-'(") = for all i < l o p . For such a string y, we have

rinalogn 5 ly l s 91ogn, 1610&910gn) < (9bgn)" i fn 2 4, and gklo8n 5 f i i f k 5 wicwn. Hence, f (z) exists. It is easy to see that f is exphonest and also polynomial-time computable.

By definition, for all k and i, [yjk-' - 4(y) c irl (y) for some y, since fifl (y) 2 fi(On).

Inspired by the notion of complexity cores, we shall introduce sets which consist of instances which occur

rareIy under most distributions in question.

Dehition 4.2.13 (PolynomiaIIy L-Rare Sets) Let C be a function on N and let 3 be a set of distribu-

tions. A set S is d e d C - m n with nspect to 3 3, for any distriiation p in F, the set jz E S I P(z) > 2-'(1'1)) is finite. If3isP-comp,t6enwesimpIydthissetSpoIynomioUyf-rme

In what foUows, we shalt show the d a c e of a polynomially &rare set for increasing, unbounded 1 with

the extra condition 5 logn 9 L(n) 5 n. The proof uses Kolmogorov complexity-

PmoE Assume that s is increasing and mtmmded In what f o h , we show a general statement: for

every distriiution p E P-comp, there exists a positive integer SIX& that, for aJl strings z of length 2 no, if z ~~[U(n) ,y(")+ ' (")e~] , then fi(x) 5 2-'(M)- This dearly implies the theorem.

For a distniution p E P-comp, using Lemma 426 , we can take a detemhbtic poiynomia-time 'Xbring

machine M such that, br all i E R fp(z) - M(x,6)1 5 2' and kM(z,O') is increasingcreasing We let T(X) = M(A,O) and T(z) = M(X,OPI+~) - ~(z-,01zf+3 for z E F. For any set A, write T(A) for LEA T(z). Notice that T may not be a distnintion, but T(z) 3 0 holds for all z. Moreover, T appra*imates fi since it fonow~ that

h m the fact that T is P-compatable, we now suppose that T on input z is computable m lzId + d steps

far some positive integer d indepemht of z.

CaAPTER 4. FEASIBLE DLSTRIBUTIONS 111

Consider an integer no Iarge enough so that the hflowing ineqdities always hold: logno 2 2 logio + 9, no 1 4c, and no 2 d + 4, where io and c are constants given later (not depending on the choice of no). For any integer n not smaller than m, let

A, = {z E 'P I T(z) + 2-"* > 2-'(")}.

As for the cardinaIity, we can daim that [[A,,[] c 2'("If'.

hf of CIoim Assmne otherwise. Note that, for each x E &, T(z) > 2-L(n) - 2-n-2 . Hence, we have

Since C(n) 5 n, CZEL fi(z) > 2 - 1 = 1. This is obviously a contradiction. Therefore, l[A,,ll < I

We note that, for each z E C", if fi(x) 5 2'C(n), then z E A. To get the desired conseqgence, we shall show that -4, C KT[2e(n),2C(n)+s(n)1~n]. Fi an integer e with 1 e < P(")+'. Let us take the eth element

of A, in the standard order on E' and consider the foIIotnring deterministic algorithm:

begin deterministic a1gorith.m

inpat (st I %, st)

let j - 0

forallstzbgsymE1do

compute the value T(y) (let this term to be t )

if t + 2-"-* 5 2-' then go to (2)

let jr=j+l

if j = e then output y

(*) end-for oatpat 0

end.

C&APTER 4. FEASIBLE DISTRlBUTIONS

since log n > 2Iogio + 10 and 5 logn I C(n) I n for aIl integers n > no. F d y 7 we shaIl analyze the nmning time of U. Notice that, m the original algorithm as defined above,

there are 2' iterations of the for-loop, and on each of such iterations, for y E Ct, the algorithm needs steps

as many as O ( T i ( y ) + logn + l + loge + 1). Therefore, the nrnning time of U on inpnt (sj,, , sf(,), sn, se)

is at most, for some constant c > 0 (not depending on the choice of n),

since 5c 5 n, t (n) 5 n, and d + 2 5 nd. Thedore, we conclude that we E KT[2(n), $(n)+'(n)l~n]. This

completes the proof. 0

Note that, by Lemma 2.7.7, the set we have used in the above theorem bdongs to the class

DTIME(O(P(")+'(~) -)I. The foIIowing Iemma of W e r [93] has the same fiavor as Theorem 4214 and will be used in Seaion 5.6.

Let k, s and C be functions on N Assume that C and s are nnbotmded and increasing- Moreover, assume

that k(n) - L(n) 5 n, C(n) 1 lOlogn, and k(n) 2 1 for almost all n. For each integer n > 0, we dehe k(n)

strings 4,. . . , 2;") as fo1Iows:

M The proof below is sitnihr to that of Theorem 42-14 For a distniution p E P-camp, take T as in the proof of Theorem 4214, This T also satisfies the ineqaality that l j i(q) - T(p) I 5 2" for all i

with 1 _< i 5 k(n). Assmne that the d u e of T(zI?), where Izl = n, is computed m time (n + m)d + d

fix some p d h e integer d We shaII de6rte co and c' later- Choose no large enough that the selfdelimiting

description of T is d c i d y small, and no > Bc', and d + 2 5 s(%).

CHAPTER 4 FEASlBLE DJSTRIBUTIONS

We shalI show that the Ipmmn also holds for aII i with k(n) 2 i > Q. Assume that the claim fails for some i

and R. Now take the minimal integer i such that ji(q) > 2-(i-a)C(n)/2 fix some n 2 n ~ , and then take the

minimal su& n. De6ne a set & as

This contradicts the m b h a k y of i. Hence, we have llrltll < 2('(n)*)/2.

To reach a contradiction, it is d c e s to show tbat Ai KT[~, 2*(9'"lz? - - - zr- Now suppose A, = {wl,rnr - . . ,to,,,), where wt < w < . - - < tom, m < 2(t(n)+4)/'. Consider the eth element we of A, where I < e < 2(C(n)*)/2. Consider the bIIowing atgorithm N which computes we:

begin deterministic algorithm for N

inpd {G 4 % G 4 Sndisuchthat lz l=( i -1)-1

let j = O

fotal ls tr ingy~z~b

compute the value T(ryP4') (Iet this term to be t )

if t + 2- < - 2-l(i-c)1/21 t h a go to (*)

let j - j + I

if j = e then output y

(*I end-fm

0- 0 end-

The aniversal lhring machine U takes an input oithe form (io,co,l(n),n,e,z,̂ -+-zF-,) and outputs we.

The size af {io, Q, Qn), n, e) is bounded by

since q 5 k(n) and [(n) 2 10 logn for all n 1 no. The nmning time of U on input ( io , @,l(n),n, e,zr -.. zZt) is at most, for some constant d > 0,

C - 2 ' ( " ) ( ~ + ( n d + d ) + [ n + 2 ) +

d - ~ ( " 1 (- + 2nd + (5 + d))

since i + 1 - co I k(n), 4d 5 n, I(n) 1 Iogn, and 4 + d 5 nd. Therefore, we condude that we E

KT[n, ~ ( ' 7 - ~ l z ? - - zEJ. 0

Suppose that we are going to solve a problem m reasonably short time w i t h the heIp of communication with

another snppIementiVjr source of infarmation. Clearly our computation depends on the accuracy of incoming

data h m the somce, Unless we can gnarantee its accmacp (eg, no cable breakdown or i n t e r f i e , etc-),

it becomes important to make our computation robust and to make it toIerate any f d t s in data. The

compntation may of course reqnire more time m case there is fadB information fiom the somce-

In N85, Schiining [W] introduced the notion of robwb muchim to model fdt-tolerant computation and

the notion of oracles he&zg such compatation. We shaIl adapt his concept to om distribution setting and

introduce a new concept, faPJt-k,Ierance of distributions, into average-case complexity theory.

For our pnrpose, we &st introduce the concept of distriiutio~~s cornputctbk relaEme to orades.

-on 42.U (Belatrvlaea * . ComputabIe Dktxiions) Let A be an orade and let C be a

-pIsrity dass-

1. A semi-distribution p is p o ~ t i r n e conrprrfoble reldiue to A (PA-computable, for short) if there

srist a poIynomiap and a- - - . oracle ' h k g machine M such that, on input (&), MA

CHAPTER 1. FEASIBLE DISTRIBUTIONS 115

works in poIynomial time and satisfies Ip(x) - MA(z, Oi)I 5 2-' for aH strings x E Em and all numbers

i E K Let PA-comp denote the set of all PA-mmputabb distributians,

2. A semi-distn'bution p is polgnmial-time amputabk relotitle to C (PC-computabIe, for short) if p is

PA-computable for some set A in C. The notation p-mmp denotes the union of all PA-comp for any

A in C.

We note, simiIar to the remark hUowing De5nition 421, that PEtomp = Ecomp and P--comp =

EXP-comp.

Dehition 4.2.17 (Fadt Tolerance) Let A be a set and C be a complexity class. A distribution p is

in P&P-comp if there exists a deterministic Turing machine M such that

(i) for every oracle 0, - M0(z, 0')I 5 2-' holds for a11 i E N and

(ii) there exists a pol~momia p satisfying ~tme&(x, 0') 5 p(lz1, i) for an z and i.

The set -4 is said to help M. Let ~ & ~ - c o r n p denote the colIection of all distributions in P~c,p-comp for

some A E C.

It is obvious that P-p Pup-comp p A a m p for every oracle A.

Proof. Assnme that p is in P&,-comp for some d There & a deterministic oracle Thing machine

M and a polynomial q which witness C( being in P&,,comp. We shall d e h e another machine M' that tries

to exhaust a11 possiile computation paths of M within time q. Here is the description of the machine

begin tun--c algorithm for M'

input (q0')

start the simnlation of M on input z

while the simulation do

if M queries y then (nondeterministidy) guess the o d e answer

exit the loop when q([zj,i) steps are consumed

end-while if M halts then output M(z) else output X

end.

Note that M' is poIpmiai-time bounded.

For each 2, Iet path, be the minimal (code of) path p of the compntation tree given by Mr on input z

such~pdoesmt~eadtotheontptttX Let n s ~ e r t h e o n t p d o f M ' o n i n p u t z a I o n g p a t h ~ ,

say &. We get b(z) - 4 I 2i became, dherwise, it violates the amdition I&) - dfC(z, Oi)l 2 2' if C

is chosen so that MC(z,Oi) = &. Nert & h e the desired set B as MOWS:

B = {(z, 1") I the ith bit of the binary ontptrt 4 of M' on input z on path path, is 1 ).

The set B belongs to because we must choose the minimal paths first, and this needs two alternations

of the existentiat state and the universal state- With the help of B as an oracle, we can compute the value MA(x,O') in time polynomial in [zj and i. Thns, p is in corn^. 0

~t is UIIICIIOWLI whether ~ P - c o r n p E P - c o m p . ow ever, o d e sets in UP n -UP do not increase

the computational power of the robust &e.

PmoE It d c e s to show that ~ ~ ~ ~ ~ - c o m ~ ~ E ~ - ~ - c o m ~ . Let p be any distriiution in

PA-comp for some A E UP n DUP, There are two polynomial-time unambiguous Thing machines No and Nt computing A and x, mspectiveIy- La M be a deterministic polynomial-time oracle Thing machine

computing p relative to A.

We use a set called a mitness of the acceptiug computation of No and Nr . Set

Witness(A) = {(x,si) I 3tu E zkI[ and No accepts x on path w whose ith bit is 1 I).

We wish to modify the machine M in order to compute the same distnibution. Let 0 = Oo $ Or , and let us

define the deterministic Turing mxrrhinr? that computes p as follows:

It is dear that N computes the distri'bution p, and if the set Witness(A) Witness(TI) is given as an

oracle, then the Rmning time of N with this d e on inpnt z is O(TiieM(z)). 0

4.3 Normalization of Semi-Distributions

In this section, we shall show how to normalize a semi-distribution to a full distribution. As the reader

can see, there are several way to normalize semi-distriiutions. Here are two simple methods: for a given

non-trivial semidiski'bution p, let

(ii) jiIr(z) = 4 fi(x), where c = limn, p(x).

Both distrtiutions p' and p" need the computab'zlitp of the limit lirn,,, p(z). This is seen as follows.

Asnrme that p is P-compnbbie (for simpLiaty, assume P(X) = 0). Assume that p' E P a m p . Let c = Cr-+X &). Smce c = lim,,, p(z), c is P-computable. Thus, by Lemma 2.7-4, the inverse 1 is also

P-computable. Since p"(z) = $ - &), p" becomes P-computable. Conversely, assume that pt' E P-comp.

Then, cis P-computable since c = ji(z) / j i"(z) for some x for which Ptt(z) > 0. Note that c = lim,, p(z) =

F(z)* T ~ w P' f P-P* One may raise the question: can dl P-compntable semi-distributions be normalized to some P-computabte

distriiutions by method (i) or @) 1. The anmm is dortttnately negative.

Proof. Take a t d y set A P which is d v e but not P-printable. Smce A is reamhe, we choose

a deterministic Turing marhme M which, on input A, produces a list of all strings of A (possibly wi th

repetition).

Now we dehe the desired semi-distribution p as

2-E if z E (0}+ and M produces 0' within 1x1 steps, B(4 =

We first claim that p is a P-computable semi*'bntion.

Set c = I&, p(z) and let O t be its b ' i representation- Notice that this represenWion is

became, for every i, the (2i + 1)th bit of r nmst be 0. Moreover,

.9smme that c is a Pampntable real number. There exists a polynomial-time lhring machine N such that IN(*) - cf I rk for .d k E K Let Nr(@) = ~ ( d H l ) . Machine N' is still p o l y n d t i m e bounded,

and AT'(@) is an initia segment ofr. Heme, fix any ~~Ecieutly Iarge n,

Thus, A is P-printable. This is a contradiction 0

We now know that we no Ionger gnaradee the existence of normaIized distriiutions. The reader may

ask under what conditions can P-computabie semidkfxi'butions normalized. In the rest of this section, we

shalI discnss a &dent condition for the normabthn of computable semidistri'butions.

First we consider a general case.

bem d i s j o i n t s e t ~ t n i t h C + = ~ ~ ~ S ~ . L e t h a n d p b e i n a r e c r s i n g ~ o n N , andpis t i m e - m l e .

Ikt Nl , . . . , Nm bc O(h(n))-time bounded ddennirdstic Ttu+g mneltineg. Assume that, for all z E E* und

k E N, tire folIoloing two conditions (i) and (ii) hold.

Then, p is O(ht(n))-Eime mmputubk, whse h'(n) = hCP(n f iIog(m) + 1) + n + ilog(m) + I).

ProoE Take pl,. . . ,p,,, and 4,. . . ,Sm, and assume sll the conditions of the lemma For the sake of

convenience, let q(n) = n + ilog(m) + 1. Consider the following aIgorithm N.

begin deterministic Turing machine N

(z,@) setResult.=o

ifz = X then go to (*)

f o r i = l t o m d o

siwtIate Ni on input (2, MC)) set Result s Result + N&, O P ( k ) )

end-for

oatput ResuIt and halt

(4 f o r i = l tomdo simuIate Ni on input ( tddk)) , W)) set Result 7 Result + ~ ( l ~ ~ k ~ ) , O q

end-for

0ntptrt1-Resultaadhatt

end,

Take a constant c > 0 such that, for aIl appropriate i, Time&, dt) I c - h(lzl+ k) and T i e q ( @ ) S c- h(k). Then the nmning time of N on input (z,@) is CacnIated as fchws for some appropriate anwants

since p and h are increasing. Hence, N is O(ht(n))-time bounded.

Next we show that N actnaIIy computes p Assume that z # A.

Forthecasez= X, we have

Consider7 for example, the standard distribution Y-. When we set p(n) = Ff - I, the distribution

v- satisfies the convergence scheme luw(l~~) - v-(lm)I < 2'i + 2-j for i, j E N.

C&4PTER 4. FEASIBLE DISTIilBUTIONS 120

Proof. Assume that p is a P-computable semi-disttl'butio~ Let M be a *time bounded deterministic M g machine computes p, where q is an appropriate po1ynod. To use Lemma 4.32, we set S = C8

and let IV(z,ok) = kf(x,OLCi) - M ( A , e l ) for dl k E N It sufficient to check conditions (i) and (ii)

in the lemma If both conditions are M e d , then Lemma 4.32 ensures the existence of the normalized

distri'bution $ that is O(q(p(n + 1) + n + 1))-time computable. S i p(n) is bounded by some poIynomiai

in n , clearly p' is P-computable.

For (i), we have

or (ii), by our assumptionT Ip(lfli)) - p ( l a ) ) l 5 2-' + 2-J for almost i, j E N. BY emm ma &4-4? this

is equivalent to the condition I lim, p(z) - ~ ( 1 f l q ) l 1 2-'. Hence? fi(S - xCJ(*)) 5 2-'. 0

4.4 Samplable Distributions

Let us consider as a simple example the generation of an LCoccttpied territory" on a finite square board

(eg, d [q). F i we randomly choose a nonnegative integer n, and dehe the "occupied territory" at

stage 0 to be the cexttm sqpare of the n x n board. At stage i, a watker randomly chooses a starting point

which is on the boundary of the board and waIks to neighboring points at random, If the waIker s t t d y

reaches an adjacent point si of the occupied t d o ry at stage i - 1, theu the territory is expanded to include

the point si. We continue to the next stage. Allowing hhikly-many stages, we are able to consider the

pmbabiIity that a certain region becomes the occtfpied territory at some stage.

This type of (probabm) distriiution is d e d swltplobZq and the algolithm which produces instances

under this distrritrtion is d e d a s~~ [9]. Instances of sampIabIe distnitrtions have o h

been o b s d in statistical physics- Samplable distriintions are also of importance in cryptography h

is known that the exkkuce of compIex samphble distrritttions Ieads to the existence of pseudcmmdom

tperato= (a [9? 4).

In 1990, Ben-Dad, Chor, GoIdreich, and Luby [9] hst formulated a notion of distributions which are

sampled (or generated) by randomized dgorithms m time poIynomiaI in the length of their output on dyadic

rational mmbers. They coined the term, potynomial s~l~phbk tW&tdhu for such distriiutions

In a recent w d c on pseud~random rmmber generators, H5stad d d 1401 aIso use an ememble of "po1yne

miat samplabI$ probdi& distxi%* To cope with dtralued dIstniuticms, we use an approodmation

scheme and give a generaIized dehitionoft-tinre J O ~ P ~ ~

Figure 4.2: A computation tree of a sampling algorithm

D a t i o n 4.4.1 (Samplable Distriintions) Let t be a function on Htt. .4 semi-distribution C( is

t-time swnplable if there exists a randomized M g machine M (which does not newssarily halt on all

computation paths), d e d a $ampling m d i m or genaator, such that

(fi(x) - R M [ M on input 0' produces x and halts within time t([x(,i)]l 2 2" (*)

for all x and i € N We say that M ~ d t n p l c s p if M d e s (*). For a set T of functions, p is T-time

santplablc if p is t-time samplable for some t E 7.

S i y , we can dehe p to be &space samphble (T-space swnplobk, resp-) by rrgniring M to be t-space

bounded (T-space bounded, resp.).

An algorithm nsed for a sam- &e is called a sampling algotithm

D M o n 4.4.2 [Q] A semi-distribution p is pdfFomiaI-time sump&b& (P-samplable, for short) if there

exists a poIynomiaI q such that p is ptiree sampIabIe Denote by P-samp the set of all P-samplable

digtniutions.

Fignre42iIInstratesacompdatiantreeofasampIing~

To distingaish our d M o n &om Ben-David, Chor, Goldreich, and Luby's [9j, we call their P-samplable

distriiutions sfriefly P-scmrplaMe. We use the notation strict-P-samp to denote the coIIection of all sttidy

P-samplable ctistn'bnti~ll~~ Notice that an &y P-sampIable dktributio~t~ are supportive.

Proof; Take an arbitrary P-samplabIe ~ ~ t x t i o n p. By definition, there exists a sampli~g machine M and a polynomial p such that, for alZ x and i,

Replace i by 1x1 in the above in*, then we have

To sixnplifj, the description, write P. for P r M [ ~ ( 0 i z I ) = z within time p([xl,lxI) 1- We then have b(x) 5 2-121 + P=. Smce the standard distribution is P-samplable, it has a its sampling machine, say &. We then dehe

the desired semi-*'bution v to be sampled by the f o l I d g sampling machine N:

begin sampling algorithm for N input X (the empty string)

choose a bit b at random

if b = 0 then simulate Mo on X and halt generate a n a t d number n ( U p generate 8,)

simulate M on input On

if M enters a halting con&uration then let x be its output and

let t be the running time of M i f t ~ p ( l z ~ , 1 x ~ tilerloatput t &output X

end.

Suppose that M outputs string z in time p([zIy 121). h the case where b = 1 and n = 121 are chosen,

for some appropriate positbe constant c, N can output x reqrtiring its computation time to be at most

c - (1x1 +p(lxl, 1x1) + I)? beeaftSe N needs O(lz1) steps to generate b a d n and needs 0@(lzl, [%I)) steps for

the siwrlation of N. Now Iet p(z) = c - (z +p(z, z) + 1) and d&e v as i(z) = RN[N(A) = z 2 q(lx[) ststeps] for dl z-

Obviously this u is a semi-distll'bution and a h P-samplable.

We fix a nonemptg string x arb'ltrarily~ Note that the probability of gen- n at random is d y

~ 2 n a ( ~ ) - 1 - Then the probabititp that n = 121 holds and N outputs z within time q([zl), is at least

Therefore* we condude that p p v.

Theonm 1.14 [9] There Qists an e f ' mumemkmumemkon of all s b i d l y P-sanplable semi-didd&hw.

In partiah, for each k > 0, them is an geclioe cnumeratitm of all &idly O(nk)-time smnphble semi-

distn'butions.

Proof, F i we effectively enumerate all randomized Turing machines (which may not halt on some com-

putation paths). Let (Mi)yN be snch an enumeration. Atso take an dective enumeration of all poIynomiaIs

with positive integer coeffidents, say @ r ) i c ~ , such that each pi(%) 2 z for aII z.

For each pair (i, j) of natural nambers, we shalt consider the ith machine Mi and the jth p o l p o d p,.

We modify the machine as follows:

begin sampling algorithm for M:ij)

input X

simulate Mi on mput X

Iet z be the output of the machine Mi and

let t be the nmning time of Mi if t 5 ~~(1x1) then output z ehe output dt-lrl

end.

Note that a random seed generated by &! is exactly the same as that generated by Mi- Suppose that Mi outputs x. The runniug time of MiQ, is O([z[ + t + 1). In the case where t 5 p,(]zl), the nmning time of

Miu, is O(pi(lzl)) because pj(n) 2 n; othemk, it is O(t). Overall, the nmning time of Mik, is at .most

c n steps m the Iength of its output-

It is easp to check that alI P-samplabk distri'buti~ appear in this enumeration. 0

The reader will find an application of Theorem 4.4.4 in Section 53.

44.2 herti'bly Samplable Distributions

Ftom a diffefed point of view, ImpagIiazm, and Levin [44] dehed "polynomiaI-time sampIable" distributions

to be of the form p p for m e p f P-comp and some f E FP. FolIowing this dehition, are can achtaiIy

construct such a distriintion that cannot belong to P-comp*

Although there is no proof separating the two notiam of "po~omiaI-time sampIabi&yP in this thesis,

we reqrrire f to be phonest, and take the f m weaker dehitiionr

1. A distriintion p is invertibly p o l ~ t i m e smplable (invettiily P-samplable, for short) if there

exists a distriiution u E P-comp and a phonest fnnction f E FP such that p = ur,. Denote by

IP-samp the set of an invertiily P-sampiable distniutions.

2. Let PI-samp be the mnection of all distriiutions of the form p,-, for a distribution u E P-comp and

a phonest, oneone function f in FP.

Roof. For the Erst incIusion, use the identity hc t ion f. Obviously f is phonest and one-one Then, we have p = prl for a11 distri'butions p. The last inclusion is trivia. 0

As shown in [Ill], in general, the feaslile computabiity of p,-I does not imply that of cr; namely, there

are distn'bntions p which are not in P-comp, bnt prr is in P-comp for the function f (z) = 01~1. Moreover,

Wang and Belanger showed that, for every p f P-comp and every increasing, phonest function f E E'P, prr belongs to P-comp [111].

We shan show the inclusion between IP-samp and P-samp.

Proof, To establish the proposition, we shan generalize the proof of Theorem 7 in [9].

Assume that u is in IP-samp. By defiaition, there exist a phonest function f E FP and a dekmhkt i c

polynomial-he Tuxhg machine M such that P = p f t and IF(=) - M(x,oi)I c 2'' for all z and i. By

Cemma 4.2.6, we can assmne that XrJbf(z,p) is nondeaeasing for each fked k- We also assume that, for

some pobomial P, I4 5 ~ ( l f ( 4 I) and If ( 4 l I p(l4 1 for x.

For simpIicitp, adte ~ ( z , @) = &-Ib) ~ ' ( z , O.(lzl)Ck-l), where Mf(z, @) = M(t. Ok) - M(x-,@).

Note that [fi(x) - Mf(x, @)I < rH1. Smce If (%)I S p(Izl), we have

To mmpIete the proof, we need to show that M(Z, @) can be computed by some sampling algorithm on

input @. La us d e h e the sampling algorithm N as foII0ws:

M(z-, ~p(lrl)+~-~) < pi 5 ~ ( z , (jp(l~l)+~-I)

if there is such an z then output f (z) and halt

end-for

end.

It is not difficult to see that ~ ( z , ok) is equal to the probability Pr[N(ok) = z in time q(Iy1, k)] for some

p o l y n o d q- 0

The converse of Proposition 4.4.7 is unlikely to hold; however, we can prove that every P-samplable

distribution is pdominated by some inverh'bly P-samplable distribution.

Lenuna 4 4.8 [I191 For mery p E P-samp, there & a d i d d n d h v E IP-samp such that jt I p v.

Proof. Let p be a P-samplabIe distniution, and let M be a randomized Turing machine witnessing p with

a time-bound polynomial p. We modify the origbl machine M so that, at every c o w o n of M, there

are exactly two nondeterministic choices. Hence, the length of a code which expresses a nondeterminiseic

path of M on each input is at least the size of its output. Moreover, we assume without Ioss of generality

that p is increasing.

We d&e a ftmction f as

output z of M on dl'l on path z' in time ~(31x1) if z = z'l and z exists,

i€z=r'lbutnosuchxexists,

if z = 2'0.

To see that f is P-computable, consider the following deterministic algorithm:

begin dete ' tic algorithm for f input r

if z = 20 then output 2 else compute z' such that z = 21

for n = 1 to lz'[

simulate M on input Cf" on path 2'

if M halts in time p(3n) then let z be its output else go to (*)

if 1x1 = n then output z and halt

(*) end-for

output X

end.

For each z, let A, be the dection of all stzings w such that, on input dlzl on the computation path

encoded by to, dd halts in time p(3[zI) and produces z. By our issumption, if w E A*, then lurl 5 p(31ZI).

[w E 4 RM[M on dlr1 produces z in time p(3IzI)I = 21w, . w

On the o t h a hand, since M appraximam j, we then have B(z) 5 2-*[4 + C I*- Let B(z) = fi-({w I f (w) = 2)). Notice that B(A) > 0. Let Q be the m;nimnl positive integer such

that Q - P(X) 1 $(A). Then, it follows that

Let q(z) = S(p(3z) + 1)* f 6. In the following, we show that q(lx1) -G(z) 2 ji(z). For z = A, this is obviously

the case- For the other strings z E P,

Unfortunately, we do not know whether P-samp S p IP-samp, or whether we can replace IP-samp in

Lemma 4.4.8 by IP1-samp.

Schder and Watanabe 1961 introduced an average version of P-samplable distnitxtions. We give a brief

definition here.

Dehi t ion 4.4.9 (Average Polynomial-T'rme SampIable Distriiutions) [98] A distn'bution p is

werage polynomial-time srrmplclble (average P-samplabIe, for short) if there exists a randomized Tming

rnarhine M and a poIpomiaI p such that

(i) - RMIM(Oi) = z] 5 Ti for arty z and i E & and

(ii) for every number r > 0 and evey ~1 E N,

Let avp-samp denote the set of all average P-samplabIe distn'bntions

We shall show below that average P-samplabilitp- is a natraal extension of P-sampIabiIitp.

Proot. Let p be any P-sampiabIe distribution- Consider a poIpomiaI p and a randomized lhring machine

M which, on mput Oi, samples p(z) m p(lz1,i) steps; that is,

CHAPTER 4. FE,4SIBBLE DLSTRlBUTIONS 127

f o r a R i ~ N a n d z E C .

To show the average P-sampIabiEty of p, we must define another sampling machine that sampIes p in

polynomial time on paverage wifh respect to the Ieugth of its output. Sucb a sampling machine is d&ed

as fouow!x

begin sampling algorithm for N inpllt 0'

simntate M on input 0'

if M reaches a halting state then do the foIIavbg:

let z be its output

let s be the random seed generated so Far

if TiieM(Oi; s) 5 p([zl, i ) then output x and halt f o r i = l t o m d o

fipafaireoin

end-for

end.

By the simnlation, it hIIows that

The rrmning time of N on input 0' dong with random seed s, is the same as that of M on 0' wi th s with

a constant factor.Thus, for an appmpriate constant c > 0, it hoI& that

which implies that p is average P-sampIab1e

4.4.3 C'osure Properties of Samplable Distributions

This snbsection w3t discnss severd properties of P-sampIable and inrertibly P-samplabIe distnittiims.

ProoE Let f be phonest and in FP- ~~e that p = v,, hr some v E P-comp and phonest g E FP.

Then,

CHAPTER 4. m4SIBLE DlSTRlBUTIONS

Proof. Assmne that p E P-samp. By Lemma 4.4.8, we can take a distribution p' fiom IP-samp such

that p SP p'. By Lemma 3.410, p S P pr impIies prl s p pkl. Now let v = p;, . By Lemma 4.4.11, we

conclude that v beIongs to IP-samp. 0

This corohry will be used to prove Propition 7.5.10-

The following theorem is similar to the fact that B P P B ~ ~ = BPP (see, eg., [1231). It states that

PBPP-samp is pegnal to P-samp. However, the proof requires complex d y s i s of a randomized algorithmthm

Roof. Since clear1y P-samp E pBPP-samp, it snffices to show that pBPP-samp EP P-simp. Consider

a distrriution p and assume that p is in pBPP-samp. There exists a M g machine M and an o d e

A E BPP such that Ifi(x) - PrM[MA(@) = 2 m time q([zl, k)]l 5 2-& for all k € & where q is an

appropriate po1ynomiaL We dehe another set At as follows. Let A' = {xlOm 1 z E A, m E N). It is easy to

see that At E BPP. Let N be a po1ynod-time probabilistic Turing machine such that

(by the Amplification Lemma). Let p(n) = 2n + 4 for all n E N Let US consider the fo11owing randomized Tbiug machine M':

begin randomized algorithm for M' input 0'

generate a natntal number at random

let cvunt F 0

forn=no tom do

while the shmkion of M on mput O'Mn)

if M qyeries y then simaIate N on mput l~(*'+P("))-lsl+p(~)

andIetcarnt~crnnrt+l

end-while

(assmnethat Mhaitsandwritesdownastringz)

if& < q(n,i +p(n)) then flip a fair coin rn times,

where m = ( q h i +p(n)) + i +p(n))(d%i + ~ (4 ) - -1 it Is1 = n then output z and balt

(*I end-for end.

Let pzi be the overall probability that z is generated by the algorithm when input 0' is given For

simplicity, write 4 = Pr,&kfA(O') = z in time &I, i)]. Remember that ji(z) = lim,, ui holds for all z.

It follows that I& - ui[ 5 2-' + 2-'.

Our goal is to show that, for any string z,

By Lemma A.4, (i) impIies the existence of the limit liq- pZi. We then set 3(z) = liw- &' . The

condition (i) impIies that

Therefore, p is p-equivalent to v.

Fix z and let n = 121- Let p i denote the probabiIity that the algorithm outpnts z after integer n is

akeady generated. This probability p: is the product of the probabilities that N correctly computes ait the

query strings made by M, and thus is at Ieast

where the kt hqgality holds by ZRmma AS. Then, we have

Let i, = &rl=Rp$ The pro- 2 is eqoivaent to the probability that, for every k 5 n, k is

fkst generated, and the alpithm fails to haIt nntil it reaches stage n of the for-loop and 6naIly wxites

z. For each k, 0 I k 5 n, with probabilitp 2-a(k)-1, the algorithm generates k it eventnany reaches

n and with probability p:, it oatputs z. Hence, if 0 5 k < n, then the probability that z is an output is

2 - a ~ ( ~ ) - l - p: - r p z ( l - e), and if k = n., then this probability is d y T-(~)-' - p';- Chera& we

have

CHAPTER 4- FEASIBLE DISTRIBUTIONS 130

The absolute value of the dXerence between 7: and 7:: is now easy to caIcnIate, = &shown below:

Using this inequality, we obtain

n-I n-k ' 2(k + I)? (27(")+1 - (2n + 1) (2-i + 2-3)) + . 2-~n)" . (2-i + 2-11

ka 2(n + 1)'

We have seen two categories offeas'bIe distriins, Pcamputable and P-samplabIe distnions, both of

which have very di&ent -cs It is natnral to raise the queseion of w M e r these notions are

truly dSrent - The ht auswer was given by Ben-David, Chor, Goldrei4 and Luby 191 who showed that

P-samp # P a m p unless NP c o w to P. Later Miltemon 1741 pointed out that P = PP is a safficieut

and mxsary condition for P u m p = P-samp, and its proof appearrd in Yasnakami [IN]- This section

shows that P-sampM.de distributions are Pamputable if and onIy if P = PP. Based on the common beGef

thatP#PP,itseemsudikeIythatPoompeqnalsP-samp,

CHAPTER 4. FEASlBLE DLSTRIBUTIONS 131

We shaa intmduce another artegwy of distnbutiorts, the so-cahd #P-cmp&dk d i s t d d i o r r p inkc+

d u d by Schder and Watanabe 1961, which seem to have more computational pawer than P-samplable

distriiutions. Again we modifg their dehition to fit our appraodmation scheme.

Definition 4.5.1 (#P-Computable Di'bufions) d [86] A distribution p is #P-computobk if

thereexistafunction f E #~andapoiynomid~rnchthat [ f i ( z ) - ~ ~ l 5 ~ ' f o r a l l z € P' a n d i ~ N.

Denote by #P-comp the set of all #P-compntabIe d i s t r i i 1 1 ~ .

We bt show that the density function ji in the above W o n can be repIaced by its distribution p-

Proof'. Assume that p is #P-mmpatable. By the definition of #Pcomputabiityt there are a set A E P

and two pdynomidr p and q such that [~ (z ) - 1 5 Ti and f (2, Oi) = Il{p I lyl = p(lz[, i ) A (I, Oi, y) E A) 11. Wdout loss of generality, we w e e that p and q are increasing.

We bt show that the function g defined as g(r,oi) = ~ ( 1 ~ 1 ~ - &= &!&$) is in #P. To see this, we

define another set A' as foIlow9:

and thus g is m #P. Now Iet g'(z,Oi) = g(z7~lzl*') and q'(n,i) = q(m,n + i + 1). Then, For any stxing z of length n, we

have:

f (x, on++') 1 Xr) - E ~ C l z l r + i + l ) --lt -*

Proposition 4.5.3 # P a p E P - c o m p = PPPcomp.

CHAPTER 4. FEASlBLE DISTRIBmONS 132

p is P-comptttabIe relative to f. Hence, #P-comp G P#Pamp. S i FP#' = fl by Lemma 2-53, it

follows that P#P-eomp = ppP-comp. u

The main theorem of this section is:

2 P-comp = #Pcomp.

3. Pcomp = P-samp.

The theorem immediately follows from the proposition and two lemmas below.

PmpmXon 4.5.5 [I191 P-samp C #P-comp.

Proof. Assume that p is P-samplable and is witnessed by a sampling algorithm M and a polynomial p.

Wrthout loss of generality, we assume that every path of M on 0' which outputs z haIts in exactly p(lz{, i )

steps. Let f ( z , 0') be the number of computation paths y such that M on 0' outpnts z and halts on path y

in time p(w , i ) . CIearIy f E #P since each path of M on 0' is bounded by p(Izl,i). It is easy to see that

the probabiIity that M(0') outputs z and hats in time p(lzl,i) equals f ( z , 0')/2~(l'1*17. Hence, p tnrns out

to be #P-computable. n

The converse inclusion, #P-comp E P-samp, is an open qnestion, The best h m result is due to

SchaIer and Watanabe [96] that every #P-compatabIe conditional distnintion can be apprmdmated within

a polynomiaI factor by some sampling algorithm in time polynomiaI m the length of outputs with nonadaptive

queries to an NP orade. This wiII be shown as Proposition 4.7.6-

The next lemma establishes a basic relationship between #P and #P-comp.

Lenuntz 4.5.6 P = PP implies P-comp = #Pcomp.

Roof. This lemma is an immediate amsequence of Propasition 4.53- However, we here &mv this lemma

in a more direct way.

Let us assume that P = PP. This is Bqrrmalient to the assfrmption FP = #P by Lemma 25.3. For

an a r b i i distn'bution p m #P-cmnp, assume that there exists a frmction f E #P and a nondeereasing

pdyn~mialpsuchthn 1jl(z)-=-l < ~ i f O r a l l r a n d i i ~ N~orwerhoathatpiscompdabIebp

some dacxmbktic 'Ihring machine in polynomial time

Define g(z, 0') = &= h(z,x, oi), where h(z, z7 0') = f (2, 0lZl+') - 2dbI*lrl+i)-dlrl*Ir[+i). Since g E #P, it foUows fiom oar assumption that g f FP. We then define the deterministic Thing machine M such that

M(z, 0') outputs g(x, 0') /2Q(~~~94, where q(n, i ) = p(n, n + i ) . Thus, M satisfies:

Hence* p E P-comp. This completes the proof.

In the foUowbg lemma, we prove that PI-samp = P-comp implies P = PP.

kmmu 1.5.7 [I191 If IPl-samp = Pcomp, then P = PP.

Proof. Let as amme that IPt-samp = P-comp: namely, for any p E P-camp and any onesne, phmest,

P-computable function f , the distniution wr is P-computabIe. We shall show that FP = #P, which is

equivalent to P = PP. Given a set A m P and a polynomial p, we set g(z) = (I{y E Hl21) I zy E .4}[1. We can assume without

Ioss of g e n e that p is strictly mmaihg. We want to show that g E FP. Now take the standard distribution v- and detine the onmne, P-computable function f as f o l k .

We also define the invertlily P-sampfabIe distn'bution q by f j = XzU-d(f-L(z)). By our assumption, r ) is

P-mmputabIe. Far the frmction g, we have the fo11owlng simple eqnatiox

where +(n) = n +p(n) + 1. Therefbre, g is P-cornpatab1e.

We comb'me the above Iemmiu and propositions to compIete the main theorem.

4.6 Universal Distributions

This section will iutroduce a slightly weaker notion of u n i d distn'bntiom, d e d prmioetsa.1 M-

bvtiorrp, and shows that there is no paniversaI distnhttion in P a m p , which is due to Sdder [95]. This

result is djfEcnIt to extend to the evert weaker notion of 00-universal distrrbtrticns.

We bqijn with the formal definitian.

Dehition 4.6.1 (7-nniversal djstriiations) Let 7 be a set of distributions and 7 a set of functions

from C to K@. A distniution p is called 7-uniaersd for T if

(i) p E T; and

(in) for every v E 3, there exists a fnnction t E 7 such that t(z) . ji(x) 2 1(z) for all strings x.

In particular, if 7 is the set of pbounded functions, then p is d e d parmersaL

The following theorem of Schuler [95] is a negative reply to the question of whether puniversd P- computable distributions exist for P-comp.

Thwnm 4.6.2 [95] Tlrere is no single P-mmputable disbibvrion which uup-dominotu dl P-computable

dkWutimOM Hence, no puniamal d i & h t h s Gdst in P-comp.

Proof. We shall show the contrapositive of the theorem. Pi we asnrme that p is P-computable and

dominates all other P-mmputable distributions. By Lemma 4.2.7(1), we can assume chat p is farther strictly

P-computable.

Using this p, we shall constrnct a P-mmpntable set which contains one string on each interval F, n E IK

We d e h e the function f fiom (0)' to E' by the foilowing procednrr:

begin deterministic aIgorithm for f input On

if n = 0 then output X

fork=ltondo

(As9amethat~=X) let L = {ale --ak-I~C"-k) and R = {@a2 - * - a ~ + r l F - ~ )

(4 iffi(L) > @(R) then let ak = 0 else let ak F 1

end-for

0UtpIlt a r e - - - &

end.

To check that f is P-computabIe is cay- By the above algarithm, f s;itisfies that [ f (0")1= n for all numbers

n E N Mareover by h e (:), the probability ji(f(On)) is at most 2-' became at each iteration ofthe f d ~ o o p ,

the probability of the set, either R or L, is reduced by halE To get the desired set, we set D = f (ID)'). Fi an a r b i i P - b i - i i e set B in E, and let A = DnB. This set A is inhite became, otherwise, the

difhxtce D - Bn D becomes an Su i te subset of z; this impIies that B is Rot P-immrme, a c o ~ c t i a n .

CHAPTER 4 FEASIBLE DBTRlBUTIONS

We next show that (A, p) E Aver(P, *).

Proof of CIaim She B E E, there exists a deterministic lhring machine Mo which computes B in time 2= + c for some positive constant c To compute A, let as dehe another &e N as follows:

begin deterministic algorithm for N

input z (say, n = 121) compute f (On) in polynomial time

if z # f (On) then *ect and halt

simulate Mo on input z and halt end.

The machine N actually computes A. NOW we must discuss the nmning time of N on input z. Suppose

that z 4 D. In this case, N needs pdynomial time. Next suppose z E D. h this case,

Hence, k.TiieN(z) is polynomial on p-average.

Let us define the dishiiution v as

6.uy(Oltl) if2 E D, C(z) =

otherwise.

Because p avpdominates v, Claim 6 implies (A, v) E Aver(P, *). To get a contradiction, we must show that

(-444 e A m p ? *I- Assmne that (A, v) € Aver(P, *), and we hail derive a contradiction- There aist a deterministic Turing

machine M and a polynomial q such that A is computabIe by M in time q on v-awrage- For every string z

in A,

T'imeaa(z) l n(lzl/~.uy(OIZ1)) 59(814(lzl+ U2)- Next we set p(z) = q(Sr(z + 1)') for aft z. Using this time bound q, the set -4 is mmitten as

A = ( x f M accepts A m time p(lz[) ).

This yields the P-comptttabiritg of R Smce A B, B is not P-immrme, a contradiction.

This completes the proot

Under the assmnption P = NP, Theorem 46.2 can be f i x r tk extended to O(f)-rmmezsat distribtttions, where f is any fanction in the set o(Zn), by a modification of the proof of Lemma 4-1 in [a].

Proof. Assume that P = NP. Assume that f E 0(2"), and p~ is O(n-universaI for P-comp. We note

that g(z) -W(z) > fia(z) hr some g E O(f) siuce p~ is O(f)-miverd Hence, W(z) 2 w- > for almost all z. Let 2-1 be the minimal string zr such that h ( z ) > 2-lfI for all z 2 i.

By de&tition, there is a poIynomiaI-time Thing m;uhine M which computes p ~ . For each z E I?, set

v(z) = M(z,@I=I~) and let i(z) = v(z) - v(z-). In general, v is not a distribution since D does not always

take a nonnegative value. Howeyer, we have fi(z) > 0 for all z 2 2-1. This is seen as fo1ows: for an z,

and thus, E(Z) 1 XI - 2-31rl > o if 2 - Now we de6ne a series of strings (zit i E N) as fobws. For convenience, write R(x, y) if y 3 21'1 and

~ ( y ) - V(Z) 2 21'1 k(y). IRt h the minimal string SU& that R(zi,zi+l) holds. This ~I&S since

otherwise, ii(z:) v(y) - I+;) < 2111 - P(y) for all y of length 2 2ifil, and thas P(y) > & for some

eanstaat c > I. For each integer n > lyj, t(y) > 2" &= = it a contradiction.

The set (zi [ i E N) is expressed by (y I 3m < ly13c0, ...,z,,, = yVi < m[zo 2 z - ~ and zi+l is the

minimal string such that R ( ~ , q + ~ ) l ) , and hence it belongs to NP. Since NP collapses to P, (zi I i E N) is in P. Note that m OD

Let tj(z) = c - 21'I(C(z) + 2-3[zl) if z E {zi[i E N); otherwise it is 0, where c is an appropriate w-tive

constant. The distribution q is obviody c o m p d l e in polynomial time, and thus, q E P-comp.

By our definition, for any constant d > 0, there c&ts an i such that

This is a con&-

4.7 Domination Relations and Equivalence Relations

As seen m Section 3.4, dominatioa relations can be viewed as an ''approJdmationn or a " r e d u c i i ~ between

two distriim*om m average-case oompkxity theory- If two distrriutions dominate each other, m this paper,

we call them "equivaId since they are dose to each other and have h o s t the same degree of mmplexiQ-

Ekpidence relations w a e &st discmsed in [%I using the terminology "appmxhation within constant

faaof to show the closeness of two conditional distnitxtions-

In this sation, we shaIl focrts on (average) po1yuomia-domination and equivalence relations and study

their properties

4.7.1 Condition f

Let us first redl that p pdominates v, denoted by v s p p, if&) -ji(z) >_ fi(z), where p is some pbotmded

fanctim Polgm- . dationsarensefulmaveragecase(pmp1exitytheorysincetheydowt

CXUPlEU 4. FEASIBLE DISTRIBUTIONS 137

change the degree of average ranning time: namely, provided that p pdominates u, if an algorithm requires

polynomial time on paverage, then this algorithm also runs in polynomial time on v-average (see Lemma

3.4.6).

Let us consider the following condition:

Condition I. Every distriiution m P-samp is pdominated by some distribution in P-comp.

The next proposition Iists severaI different conditions which are equivalent to Condition I,

PrapodiCion 1.7.1 The fotlorping amdiths ore equivalent.

I . For every p E P-samp, then ezis& a W b u t i o n v in P-comp such that p SP v.

Proof. Smce IP-samp C P-samp, (1) implies (2). By Lemma 3.4.13(1), (2) is equivalent to (3). We show

that (2) impIies (I). Assume (2). For every p E P-samp, take a p' f Ip-samp which pdominates p by

Lemma 4.48. Use our assumption to obtain a distriiution u E P-comp which pdominates pt- By Lemma

3.45(1), we have p v. 0

By Theorem 4.5.4, Condition I is derived from the assumption P = PP. Ben-David, Chor, Goldreieh,

and Luby [9j further shuw that if Cbndition I holds, then no strong one-way function exists. The foIIowing

is an important fiapent of their prooE

(iii) M o n t n p u t z ~ l i s f s d J e m e n t s 0 f f - ~ ( x ) ( w h e n e o ~ ~ f - ' ( z ) = 0 , M o u P p u t s O ) i n p o l y n o m i a l

Eime unless z E S.

ProoE Now let h({w, 2)) be (y, z) if w E {O)', It/ = [tal, z E x, and f (z) = y. Note that h is d d e h e d and phonest, Take a distriiution p detined as f m . @((w,z)) =-B-(z) - 2*-*(l~l)-~ if w E {O}rT or

else 0- CI-ly, phl E IP-p c Psamp Note that k ( 2 ) > *=. By our assmnption Condition I (as well as Proposition 4.7-1(2)), there are an q E P-comp and a poIyne

midrsuchthrrr(~)-$(*.I) >fihqb,%) krdlyandr- !%, breach)~ran(f),wehavefj(~,z) L $i$ fix some r, where s(n) = 4r(n]nk (n + n' + For each g, let C, = {z I fjb, z) 2 h, r,lr[ <_ 131 + klog lgl}.

Define S = {y I l[C,l] > q(lul)s(ly[), y € ran(f)). Clearly we have S G ran(f). We show that, for almost aIl n, [IS11 < &. Assume otherwise, and kt I/ be an element of S. la n = ly'l. Then we have

a contradiction. Define a k i n g machine M as foil- on input y, by a depth-first search, M computes alI

elements of C. if llC,ll < q(lyI)s(lpl) and lists aR elements z of C, which satisfy f (z) = y, or else M ontpats

0. Since f"(y) C', all elements of f-l(y) are printable in polynomial time if y E 3. This completes the

proof.

Using the bash-function technique of [MI and the ampIification technique of probabiic W g machines,

we on show that Condition I Ieads to the consequence that every N P set is nearIy-RP.

Theorem 4.7.9 [119] Asswne Condition I. Let A be OIQ set in NP. For eany polynomial p with

p(n) 2 I f o r d l n E N, there ezist a set D a d apol3pwmdal-time nnufomiz~ lbingrnaclaine M such that

D A, and, for each z, z E A- D implies RM [M (2) # A(z)] < 4, z 4 A implies PrM [M (2) # A(z)] = 0,

and Pr,,[z E Dl < &J fw almost all n. H- Condition I implies thut mery NP set i s nead*.

Proof'. Assume Condition I Take any set A in NP and any poIynomial q, and we wil l show that -4

satisfies the claim. There exists a set B E P such that A = {z 1 32 E ~ l ~ l [ z z E B]} . Let B, = {z I zz E B )

for each z. Assmne that there exists an increasing polynomial p such that Pr,[z E A] 3 &J for almost aI1

n since, otherwise, the theorem is trivial by &&g D = 0.

We take the set H%n+, of hash fanctiom. Let $ be the kth string in the set x 3 ~ ( " ) with respect to

the standard order. Dehe f (2) = l~$hh(y) ,+&~-~ if z' = zysEh and y E B,; otherwise W, where z € P, h E H,,*, and c = iIog(n). Notice that Is'] 5 If (d)I for all d, For each z, let g(x) = IIlf-'(z)ll-

For brevity, write t(n) = t + n + iIog(n) + (n + I)(n + ibg(n)) + n + do&).

For fixed k, and z of length n, let p b = Rhm[9(k4h~M+E~"-L) = 1 I h E Hn,n.tc, w E P*]. We

k t show that pkJ > & for almost all n. Now we 6.x k and 2, and assume that n = lzl is dcientIy

large and iIog(g(z)) 5 k 5 n. Consider the case iBzI[ > 0. The probability p ~ , is larger than or equal to

the probability over all hw that, for each y m B,, h(y)+& = w + ~ , and h k-distingrdshes y on Bz- Thus,

since logn i c <_ Iogn + 1, For the case [jBza = 0, cIearip p q = 1 since g(Iz$inatn+EO"-k) = I for dl k,

h, and w. This yieIds the desired result,

By Lemma 472, t h e are a set S and a pdynod-time ddeterministic Tming m;uhinn N which m- y(=)

ognkes 3 mch that S r a n 0 and IIS n fin)[ < . We dehe a randomized poIynomia-time

algorithm M as fonows

CHAPTER 4. FEASIBLE DISTRIBUTIONS

choose w and h at random (tu E P*, h E Hn,n+c, c = ilog(n))

let Result = 0

for all k (1 5 k 5 n)

nlll N On dk = ~ S ~ ~ W + ~ O " - ~

let Resuft F OR of Result and N(dk)

end-for

output Result and halt

end.

Cet &, = R h m [ l ~ $ l n r r + ~ ~ ~ ' ~ € S I w E IF*, h E Hn,nk], where c = ilag(n). Using this '&,,

we de6ne D = {z E Cf n A I 3k[irog(llBrII) 5 k 5 n A b r , 2 A 1x1 = n]). We wiI l show that

& n [ ~ E D] < A. Assume otherwise So, we have & > & for some k (dog(ll BJ) I k I n) and

z E Dn. Since

we have max(&, I iIog(llB,II) 5 k 5 n,z E P) < &. This is a contradiction. Therefore, Pr,[z E

Dl < A- Now our goal is to prove that (i) RM [M(z) = A(z) J > for all z in A - D, and (ii) &[M (2) #

A(z)] = 0 for all x $l A. This is enough to establish the theorem because of the worst-case version of

the Amplification Lemma Take any input z of Ieagth n. Let p, = Rh,[A(x) = ORL, N ( 4 ) ) [ h E

Hw,w E P*]. Note that the probabilitp P~M[M(x) = A(z)] is at least p,. .Qsslrme A(z) = 1 fot a

string z E n. Note that if dp E 3 and g(dF) = 1 for some K, then ORL, N(z-~) = 1. Hence,

For the other case A(x) = 0, N ( Z @ - ~ ~ ~ O ~ - ~ ) = 0 for all h, ut, and k; and thus, p, = 2. This compIetes

the proof. n

4.7.2 Condition P

PoIynomiaIdomination relations are useful but too tight to be considered an &ective measnre of "approx-

imation" or "reduabilitf between &sttibutions in a- complexity theory- Gurevich [36] h e r

introduced a weaker farm af domination relations by tegniring a frmction to be p-bounded "on the average"

FoHowing his dehition, we a h Condition I to allow the domination to be, instead, "polynomially-

bounded on the avemge+n

Condition It. Every distribution in P-samp is avpdominated by some distriiution m P-comp.

Obvioasty Condition I irnpIies Condition It, but we suspect that the the umverse impJicatioa may not be

provable afbmtidy.

In the case of avpdominntions, we no 10- prove a similar equivalence as in Proposition 4.7.1. The

fo110wing claim is the best possl'b1e so far.

Proof. By Lemma 3.4.13(2). 0

To see a collSeQtlence of Condition It, we need a notion of Ph, and its strnctnral properties- These

wil I appear in Section 7.5-

We & see the gap between the two notions: #P-camputabiIitp and P-samplability- The next proposition

was shown by Schuler and Watanabe [96] for conditiod distributions- We modifp their proof to accomme

date inhihe distriiutiom. To descni the proposition, we need the notion of *table samplabii of

distnLbutions*

1. A distriiution p is ~ & s m n ~ ~ o ~ c if there exist a randomized o d e ntring machine M, a deterministic

'Ilning machine N, and a pdGomiaI p such that

(i) M with orade A generates p m time poIynomial m the Iength of outputs; and

(ii) on input (Oi, s), N lists all query strings of Q(M, A, Oi, s) m polynomial-time (without any queries)

if s is a code of a computation path given by M wi th orade A on input Oi.

2 The notation Pi-samp derides the co1Iection of all P&axnPMIe distriintio~~~, and Pen-samP denotes

the anion of alI distributions in Pt-samp for any A E C.

ProoE Tahe an arbitrary #P-computabIe distnibntion p- There Bdst a frmdon f E #P and a polynomial

q such tbat [fi(z) - [ 5 2i for an z a d i E N Wrthout loss of g e n e , we can assttrue that this q

isinaxsting. ~asirn~~icitp.,let$,=~~~fmeachi~~ since f €#P,thereexistsasetA~Psuchthat

f (2, 0') = l{w E ~dl*l*) I (06,x, w) E A}[ for dI 2. For brevity, let & = {(w, z) I m E A (Oi,x, w) E

A). Let$ ={am1 (x,w} € & A l z l = n A k l =p(n,i)). Noteby Lenma A-4 that /<-oil sr'+?-j forallidegersi,j>O.

CHAPTER 4. FE.4SIBL.E DLSTRlBUTIONS 141

We use u n i d hash functions to appmxhate the value d w. Let us consider the set

Hn+4(Ri),n+o(Ri7+l of hash frmdion~. We dehe the following three sets that will be used as oracles:

begin randomized algorithm M input 0'

generate a natural number no at random

for n = no to oo do

if ( ~ ~ + p ( ~ ) , on) 4 Xo then go to (**)

repeat i + p(n) times

generate h in Hm,m+l at random, where m = n + q(n, i + p(n))

generate y in P+' at random

if (0'+P(~),0",h,y) E XI then go to (*)

6nd the strings z E FF and w E CQ("'+p(")) us@ X2 st.

h(m), = y,, and (2, w} E Ai, where r = q(n,i +p(n))

output z and halt

(*I end-repeat

(**I end-fm end.

For each z, we W e F,' as the probability that the randomized algorithm M outputs z. As m the proof

of Theorem 44.13, it d c e s to show that

(ii) pXi - p,J[ 5 2' + 2-3 for all i, j > 0.

To get the desired distribution v, we set O(z) = I&- a', and amseqnentIy, p is p-equhdent to v.

In the fo1Iowiag, we shall prove cIahs (i) and (i) above E l i i and n- Let m = n + q(n,i) and

r = q(n,i +p(n)). Moreover, Iet !$ = {w 1 (qw) E &,)). Fm we d& & as the pmbabihy, over

aU h E &,,,+I a d aIl y E C"+I, that the algorith M h d s z in a single iteration of the repeat-Imp;

d y 7

This is equivalent to the following

Hence? by Proposition 262(2), we have (retting c = 1 in the proposition)

Next we define p i as the pmbab*, over all h h HB,,+l and all y E F+', that the algorithm M hds s

dnring i + p(n) iterations of the repeat-Imp. We 6st estimate the error probability. The error probability

is at most 2-i-p(n). As a consequence, we have

As in Theorem 4.4.13, we can show that Ip: -dl 5 2-~(~)+'(2-' + 2-9.

The following analysis of the algorithm is similar to the proof of Theorem 4.4.13.

Let 2 = The probabdity pZi is equivalent to the probability that, for every k 5 n, k is

6rst generated, and the dgodthm faiIs to halt until it reaches stage n of the for-loop and finds z. For each

I , with probability hr we generate integer k. FolIowing the algorithm, we eventually reach n and

with probabilitp p',, we ontpnt z. Hence, if 0 k < n, then the probability that z is actuaJly outpat is 2-*(k)-1 - p; r[$(l- $1, and if k = n, then this pmbabi i is exactly 2-m0g(n)-' . p i . Thus,

Hence, we have

For the condition (ii), we fdIow the proof of Proposition 4-4-13. 0

There may be possibIe to replace the sgmbor T'' m the above proposition by "CF However, because

our algorithm needs to know the Iength of oatput shings z before the actaaI simaktion of P(z), we cannot

condude that #P-comp G F - s a m p

Here is a ooroIlarp of Proposition 4.7.6. We note that NP BPP if and only if NP = RP [521.

Cod- 4.7.7 NP BPP iinplics #Pamp 9 P-samp.

RooE We note that P-samp E #Psomp. It saffices to show that #Psomp EP P-samp under the

assamption NP E BPP. Let us assume NP E BPP. Take an arbitrary distribution p in #P-amp. By

Proposition 4-76, there is a v E pY-samp such that p is Peqtdpalent to v. Under our assumption, v

belong to ~ ~ ~ ~ - s a m p . Using Proposition 44.13, there is a distn'bution { in P-samp sa& that v np <. Hence, p =P <. I3

4.7.4 Condition II'

In this subsection, we study the foIIowing condition on equivalence relations:

Condition II'. P-samp caw P-comp.

CIear1y Condition II' b p k Condition It. As m the previous subsection, we can show the following proposition.

Proof. ByLemma3.4.14(2). 0

In what foIIows7 we shan show that IP1-samp cav P-eomp implies P = RP, and thus if Condition

II' is true, then Re cokpses to P. In the totlowing proof, the worst-case d o n of the AmpIihtion

Lemms to one-sided bounded-error probabMc algorithms is &ectmely used to make its error probability

exponentianyd 0

P l p p o ~ 4.7.9 [IlS] IP1-samp c- P a m p i r n p b b P = RP.

Root Consider an arb'rtrary A E RP. We prove that A belongs to P. By the amplification lemma [91],

there is a skidy increashg polynomial p and a set B E P such that, for every x E F, Pr,[(x7 g) 6 B [ y E

5 2- if x E A, and otherwise, R&, y) E B I y E = 0.

Let p be the distn'bution d&ed by fi(ly) = fi-(z) - 2-P(lZl) if ly[= ~ ( 1 ~ 1 ) ~ or else i(zp) = 0. CIearIy

/r is P-comptrtabIe- Let

CK4PTEU 4. FEASIBLE DISTRIBUTIONS 144

where r(n) = (2n +p(n))l. Since u E Pcomp, there exists a deterministic polynomial-time M g &e

M such that Ifi(z) - M(z,Oi)l < 2-'. Let M'(z) = ~(z,0~(l~l)+~l'1). By definition, It(z) - Mr(z)l c 2-f(l.+l)-2lrl for 2.

Let z E C. Assume that z E A. Then, we have

since V-&) 2 & 2 & if n 3 7. Hence, M'(z) > fi(zlP(")) - 2-r(n)'2n 3 ~ f ( " ) - * ( 2 ~ - 2). In D, q(z) fi(z) = 0. Hence, Mr(z) < fi(zlp(")) + 2'r(n)-k = the case that z B A, fi(z1dn)) I Cze f - t ( r l ~

2-fln)-&. Now we have a complete c k a c m k t i o n of A in terms of Mf; namely, A n C = (z E C (

Mr(z) 2 2-r(n)-2n(2n - 2)) for almost all n. S ice M' halts in polynomial time, A is aIso computable in

polynomial time. D

The above proposition does not d i c e to imply that P = PP, since the worst-case d o n of the

AmpIification Lemma may not hoId for PP sets.

4.7.5 Condition I1

in contrast to Condition Ll', we shdl consider its polynomial --on, which leads to the mnclasion that

P = NP. Formany we dehe Condition II as f o U m

Condition IL P-samp S p P-comp.

Note that Condition 11 implies Gmdition II' as welI as Condition I and that, by Theom 45.4, Condition

IIis trueif FP =#P. We start wi th the following proposition.

CHAPTER 4. FEASIBLEDISTRIBUTIONS 145

Proof. By lemma 3.4.14(1). 0

B y Proposith 4-7-10, we can replace P-samp in Condition II by a d e r set IP-samp. At the end of

this section,weshallseet&t wealsocanrepIaceP-samp by alargerset #P-comp.

Next we shall prove that Condition II yields the consequence that NP collapses to P. We first strengthen

Lemma 4.7.2 under the assumption that P-samp is pincluded in P-comp.

Lemma 47.11 [I191 Assume that P-samp EP P-eomp- For m y set B E P and any polynomial p, let

SB = {Z I IIBJ <p( l z l ) ) , mrhereB, = {z E ~ 1 ~ 1 I xz E B). There exists a detffmmutre * . .

%ing machine M such that, f m e a c h n ~ N, M on inputt inSgnP listsdlckmentsofB, (whemuerB,=0, M outputs

0) in polynomial time.

Proof. Assume that P-samp CP P-comp. We define a phonest function f as follows: f ((w, yz)) = (ly, z)

if w = s p l for some k, [zl = 191, Izl = k, z & z, and y2 E B; otherwise, (Oy,z). Let fi((ur,z)) =

Vstolrd (z) 2- 21108(Izl)-~ if w E ~ ~ q ( l ~ l ) , or else 0, where no&n) = llog o no&).

Since p,-z E P-samp, our assumption ensures that there are an q E P-comp and a poIynomiaI r such that

r(lyl+Izl).fi,-t((y,z)) 3 i((y,z)) 3 fip((y,z))/r(lyl+lzl) for ally and z. Denote by D, the collection ofz 2-at#[

s d that 2 E land E Br If y E Ss, then llDlill 5 ~(1~1)- Note that fif l((y7z)) = ~&.-2u*nn.l,+~ 9-21.1

For h p h d t p , let q(n) be Zllog'(n) + 2110dh) + 2- Hence, ri((ll, 2)) 2 r(l jl.mlyl, 8 DI. # 0. Let N be a polynomiaI-time Turing machine which computes ij. Let d be the minimal positive in-

teger such that 3 - t(2n) - 29(") 5 2d.no6(n) for almost all n. We define a new machine N' as Nr((y,z)) = ~( (p , z), @lH.uOdlpl)+f ). Hence, Nf((ly, r)) > ri((lg, 2)) - T ~ ~ ~ ~ - ~ - ~ O ~ J J T ~ ) - ' > s- For eK6 y E p,

2.2'" let C, = (2 I N'((~Y, 4) > -1 I4 5 n). 2-2Iv1

We note that if r E C,, then fj((ly,x)) > wi. For each n > 0 and any y E SB n C,

Therefore, IIC,[ 5 Pwn)-0(") - r(!h)* . C1 11D3,-1[ 5 2%(2n)2p(n)3n.

Note that B, E C, if y E SB. Smce C, is printable in poIymmial time, all elements of B, are printable

in po1ynomial time. This completes the proof. U

ProoE h us assame that P-samp P-comp. Let A be an a r b i i set m NP. we shall show that

A E RP sinceP =RP byPsoposition4-73. It d c e s to consideraset A d t h e f o r m A = {x [ 3 E

!F.lx1[~€ ~ 1 ) f o r s o r n e B ~ ~ - Let B , = ( X E X M IzzEB) .

Letnsd*

CHAPTER 4- FEASIBLE DISTRIBUTIONS

Sice B E P, B is also in P. Let Sg = {d ( [&(f I), where Be = { i E zI"~ ( ii E b). We define pk, = P r ~ [ z s ~ h w + + ~ * ' ~ E Sg = 1 1 h E H%,+, A w E Cn+q. We first show that, for

almost a11 n and all z with (IB,I( > 0, pk, > &. Now 6x k and z and assume that n = [zl is sufficiently

large, and ilog(((B,(I) 5 k I n. The probabilitg pkT is at most the sum of the probabilities over all strings

hw that, for each z E B,, h(z)+- = w++, and h k-distingrrishes z on B,. Thus, we have

We apply Lemma 4.7.11 to the set Sh, and we obtain a polynomial-time deterministic Trrring machine

N which recognize Sg. We define a randomized pdynomid-time algorithm M as foUows:

begin randomized algorithm hi

input z (say, n = Izf)

choose w, h at random (w E P, h E IThnfc, c = ilog(n))

let Resutt = 0

f m a l l k ( l < k < n )

ram N on dk = 6 h . p

let R d t = OR of ResJt and N ( 4 )

end-for

oatpat Result

Oar goal is to prove that R M [ M ( x ) = A(z)I 2 6 br almost all x. Take any input z of Iength n.

Let p, = Pm,[A(z) = ORL,N( ik ) ) ( h E Hwk, w E PI. Note that the probability PrM[M(z ) = A(z)] is at Ieagt &.

A(z) = 1. Note that if 0 c gb(dp) < 1 fm some K, then ORL1N( ik ) = 1. The probability

pt = P f k [ A ( x ) = ORLIN(z$) h f H,,*, w E Cn] is at least the smn of the probability over an htu that, for each z E Bz and for some k with IIog(gB(z)) I k I n, h(z)+- = W+H, and h k-distinguishes

z on B,. Hence, p, > gs(z) ( I - F) -Z(W 2 (1 - r E ) Z e 2 For the other case A(z) = 0,

N ( ~ $ I n o ~ 0 " ' ~ ) = 0 for a11 h, w, and k; thns p, = I. This completes the proof. 0

Prod IIf - part) This is obvious a c e #f-comp 1 P-samp by Proposition 45.5-

( M y if- part) Assume Conditi011 IL By Theomin 47.12, we have P = NP- In particnlar, NF _C BPP. By Corolbtp h7.7, evexy #P-a>mpntabIe d i s t r i i is pegnmalent to some distriitrtion which can be

sampled by a randomized 'Jhting machine in time polynomial in its output

4.8 Other Topics

There are several intriguing distniutional issues which are not discussed in this chapter. Here we present

two diffetent approaches.

Ranking Distrr'btttions. Reis&& and Schindehauer [84l introduced a new type of distributioxs that

dows precise compIdty drrcsification of distriiutional problems. They called such diseibutions mhbk

G i a distribution p, its m n b g jkctirn rankP(z) is defined as rank&) = 11 {I I P(z) 2 fi(z)}11. A distriintion p is p o E y n d I l g mr3cbfe (P-fankable, for short) if the fnnction rankP is oneone and P-

mmputab1e [84]. Let P-rank denote the collection of all P-rankable distributions. Note that, for every z,

raakp(z) - /i(x) c, B(z) = 1- This implies that rankP(x) 5 & unless B(z) = 0. In particular, if p is

positive and supportivet log(rax&,(z)) 5 p(lz1) holds for some polynomial p.

For simplicityt let t be a strictly increasing fanction on P. A function f from X* to N is called t-menye

with respect to rankP if x. ,,,,, c+ 5 rn for any number m 2 0. We say that a dkiiution

(D, rankP) is solvable in maoge p v 2 ~ time with Rspcd to mnkaMQ if there exist a p o l y n o ~ p and

a det ' ' t ic 'I\lring machine M computing D such that h.TiieM(z) is gaverage with respect to dP.

Watanabe [I141 pointed out the foIIowing relationship between Psomputable distributions and P- ranlgble distributions. Take a distriiution p satisfping log(I.ankp(x)) 5 [zld for some constant d > O.

Roof Sketch, Assume that f is polynomial on &-average Then we can take constants c, k > 0 such

that, for a11 m > 0,

For any m, let X, be the set of strings z such that T L - I < rankP(z) 5 P. Now we can daim that

there exists a constant e > 0 such that, for any sufTiaently Iarge m,

To show the daim, let X'& = {z E X, I f (z)'fk/W 5 m4) and XE = { x E X, I f (z)'fk/[d > m4). For

On the other hand, the set Xk &i&es that

We note that Belanger and Wang [6] show that X F W P is in Aver(P, *), then any problem in Dist(NP, P-rank)

is solvab1e in average po1ynomial time with respect to rartkabiky.

IndistingaisbabiLity of Distri'bations. Another topic is the complex?y of &triibutioxts. We &dl

f o ~ m Meyet's termino1ogy [q. Two clistriions p and v are called %tatistically indistinguishabIen if

lim,- nk ,,=,, IA(z) - = 0 for d 6 E N p and v are "circuit indistingnishabten it,

for a~ f;unities of polynomial-size circuits C = {Cn}n.N, Lim,- nkl Cn(z)&(x) - A(z))l = o for aD k E & a d uaIgorithmicaIly mdistingnishablen if, for dl polynomial-time probabilistic 'Ihrhrg machines

M, ~im,- nkI & , l , R ~ [ M ( ~ ) = I] - &(z) - S(z))l = 0 for dl k E N Meyer [73] show the foIIowing separation d t :

We do not intend to indude any p d of this theorem. The interested reader may refer to [n].

Chapter 5

Average Polynomial Time

Reducibilit ies

5.1 Introduction

In the early 1970'~~ Cook [a], Karp [49], and Levin [59] took a pioneering step towards the dadication

of the hardest problems in JW. They introduced notions of resourcebounded reduabiies amrmg NP

problems. The idea of these reducibilities is as foflows a problem T is recognized to be at least as lrmd

as auother problem S if the prob1em S can be transformed into the problem T m polynomial time so that

if T is solved easily, then so is S. The problems "hardestn m this sense among NP problems are d e d

NP-d~npkte TypicaI NPuuupIete problem are s- problem, Emniltonicm eireuit probiem,

and trno* sateman problem For more NP-complete problems, see [26J.

In averagecase complexity theory, Levin bas introduced a similar notion of polynomial-time many-one

reduciiihty among d i s t n i a t i d decision probIems and showed that the randomized bottuded tiling problem

is one of the hardest problems m Dist(NP, P-comp). His reduction from a dis tr i iut id problem (-4, p) to

another distn'butional problem (B, u) reqah a polynomial-time many-one reduction that maps set A to set

B and employs a so-ded dorninaEion condifion betwen p and v. This domination condition gnarantees

that instances occuning with high probability are mapped by the reduction to instances oecnrring with

high probabw. This reqnirement is d in Levin's theory of amqpase NP-compIeteness in otder

to ertsme the dosure property of a cIass under these reductions- At the same time, it makes completeness

prooh of given distributional problems diEcnlt to toeve.

SiLevin'sdiscloverpofa~NP-completeness, researchershaveintroducedseveralinteresting

notions of reducibilities into the theory of awmpxse NP-completeness- We will review these redncibitities

m this chapter.

Seaion 5 2 will introduce two ofthe most important reducibiIities among distributional decision pbIems:

determiaistic mmy-one reducibility and TWng reducibihy- Levin [60j introduced a notim of mmy-one

reduciiility, and the notions of deterministic 'Ihring reduablitp and mdom many-one reducibility were

introduced and studied in [60,9,44,12,96]. Other reductions of interest are logspace maLly-one reductions

[9] and logspace many-one reductions which are phonest [321.

Section 52 will introduce detamirristic many-one redtlcibility and po lpmi - t ime is~llorphim among

distn'butiond decision problems, and then the notion of deterministic ZWng reductbility will be introduced.

The Thing reducib'i wiIl be partidarly used to build the average polynomial-time hierarchy in Chapter 6.

In Section 53, we shall &Ti severaI many-ate complete problems for Dist(NP,P-comp). We first

show that the randomiEed &mdd haking p b h is comp1ete for Dist(NP,Pcomp), and then the other

distn'butional problems are shown to be po Iynody isomorphic to this problem.

In Section 5.4, we shall discass several incompleteness d t s , The choice of distributions is important

when one attempts to prove that some distnLbutional probkm is complete for Dist(NP,Pcomp). Some

distriitltions are such that their associated distriiukional problems m o t complete for Dist(NP, P-comp).

For exampIe, fiat distrr'butions aud sparse distributions have this chackm& . . c.

Flat distributions are often arise in graph-dated decision prob1ems; however, no distributional prob1e~1~

with h t distriidions become complete for Dist(NP,P-comp) tmless EXP = NEXP. V d t e s a n and

kvin [I061 proposed new reduction which uses random coin tosses. In Section 5.5, we shaII mintduce

probabilistic reducibility, d e d botcnded-enw prr,ba&Wk trufA-trrbk dvcibility.

In Section 5.6, sone structural d t s are shown.

Major Contrr'butions. Some d t s in Section 5.6 are from Wt11e.r and Yamakami [97l, and Section 5.5

provides a series of new resnltts. More precisely, the following are the major contn'butions to this chapter.

Theorem 5.45 shows that sparse distriiutions fail to make the corresponding distri'btttional problems

pm-compIete for Dist(NP, P-comp) dess P = NP. Lemma 5.5.5 Is a r e l a th id version of the AmpEcation Lemrna (Lemma 3.5.31).

Lemma 55.4 shows that skew avbpptt-rednaiititg is closed under pm-reductions; namely, if (A, p) is

pm-reduaIble to some distn3utional problem which is skew avbpptt-reduui1e to (B, v) , then (A, p) is skew

avbpptt-reduaile to (B, v) . Lemma 5-52 shows that, if (A, p) is skew avbpptt-reduciiile to (B, u), then there exists a pmbIern (B', J)

which is pmredtxu'b1e to (B, v) such that (A, p) is skew avbpptGredu~'b1e to (B', r / ) via a reduction machiae

which queries strings of length greater than the input size.

Proposition 5.6.5 shows that, for every recnrsive set D not m P, them exists an incomparable pairs with

respea to s. Theorem 5.53 shows that the diskiiutiond bounded halting probIem with a fIat distniution is bpptt-

comp1ete for Dist(NP, Pump).

Proposition 5.5-7 introduces a new d t regarding the transitivity of avbpptt- and bpptt-redua'bilities-

5.2 Deterministic Reducibility

Reduciiility is one of the most important took m computatiord complexity theory for assessing the d-

ative comp1- of two given problems, This section wiIl present the formal definitions of deterministic

reduciiilities arnong distributionat decision problems and develop their basic s t r a c t t d properties.

5.2.1 Many-One Reduu'bility

In this section, we shall formally introduce two notions ofmarty-one reduci%Sties: potpMILial-time many-one

and auerrrge potynomicJ-m many-one reduciiihies.

A many-one reducibility among distributional decision problems requires mappings among sets and con-

straints among distniutioons. Let us consider the worst-case polynomial-time many-one reduction f between

two sets A and B. The reduction ensares the relationship A = {z I f (x ) E B). If B is computed by a

machine M in polynomial time, then a simple algorithm N which fiRt computes f(z) and then simuIates

machine M on input f(z) amtany witnesses the P-computability of A. Now we switch h m sets A and B

to two distributional decision problems (A, p) and (B, v). Suppose that (By v) is computed by a machine M in polynomial time on u-average; that is, Az.TiieM(z) is polynomisl on u-average The above dgorithm N

needs at most O(Tmef (2) + TimeM (f (z))) steps, where Timef (x) denotes the time teqnired to compute

f (z). To g~arantee the average polynomial-time comptrtabiIity of (.& p), the dues i(2) and P(f (2)) must

be dosely related This Iast requirement is called a "domination condition" for f.

Fm recdl that the notation vrl denotes the d M t discr'bufion defined kom u and f by its probabiliw

(2) = P({r I f (2) = 2)) for each z.

There are several possr'bilities for domination conditions. Here is a list of popular conditions:

(ii) C(%) 2 zw(p) # for some pbotmded function p.

By Lemma 3.411, condition (in) is equivalent to the condition that p S P 7 and u 2 r)/-~ for some semi-

distribution q, and thnq @) implies (5). By the pmof of Lemma 3.413, (i) implies @) if f is pbotmded

Assume that f is phonest and pbotmded- By Lemma 3-4.12, (ii) implies (iii), and by Lemma 3-4-13, (Z)

implies (i).

In most cases, the choice of domination condition is relatively harmless when one proves atrerage NP- completeness of a given distn'btxtional decision pmbIem, beeatlse most known complete problems are isomor-

phic. In this thesis, however, we shalt use the weakest daminattim condition (ii).

De_&lition 5.2.1 (PoIynodakTime Many-One Reductions) [q Given two distn'butiond decision

pmbIem~ (A, p) and (B, v), (A, p) is edkd p d p m i d - E i r n e nurngonc M l e (p-m-reduu%1e, for short)

to (B, v), denoted by (A, p) Ip, (B, v), if there ez&s a firncticm f such that

(ii) (Domination) p * r), and P 2 for some semi-distn'bution q.

The hc t ion f is d e d a poled-Eime my-one r e d v d h @-m-reduction, for short) jbtcth and is

said to reduce (A,p) to (B,v).

Gurevich [3q expanded many-one redua%iEty by dowing a many-one reduction to be polynomial on the

average. This reduction is known as memge polynonrid-time manpone red~bi l i ty-

Dehition 5.23 (Average Polynomiall-Time Many-One Redo&*ons) [36] Let (A,p) and (B, u)

be two distributional decision problems. (A+) is mmge polyrwmid-time mapone reducible (avpm-

reducibIe, for short) to (B, u), denoted by (A, p) (B, v), if there exists a hc t ion f such that

(ii) (VaIidity) A = (x I f (2) E B); and

(6) (Domination) p 5- q, and P 2 ijrl for some semi-distriiution q.

The condition (iii) on the distriitrtions m the above dehition is simply called the dmninahbn d o n

for the reduction fimction f.

Om definition basicatty foIIows Gorevich's domination condition. Gurevich [36] d&ed his many-one

rednability as follows: (A+) pdynomiaIIy rnatly-one reducible to (B,u) if and only if there erdsts a P-

computab1e function f and a distn'btrtion r) and a set C C_ {z I P(z) > O) such that f reduces An {z I b(z) > 0) to B, p SP I), and GrC(y) = +(y) for dl y, where urc is the &tri'bution satistying that

f

The fo11owing lemma is -0-

Proof. Let us consider the id- fimction f, ie-, f(z) = x for ail z. The assumption p s p v is

eqnivalent to fir, P v- Hence, (D, p) is pm-reduci'bIe to (D, v ) via f. a

The fdowhg lemma is a d tod for proving that the composition of fimctiom is polynomial on

P-average

CHAPTER 5- AVERAGE POLYNOMIAL ZlME REDUCIBEJTES 153

Roof. .4snre that p 5- q and P 2 for some semi-distribution rl. Asmme that f is pf on

waverage, and )LZtJg(z)I is pg on paverage* where pf and p, are appropriate increasing polynomkk Sice

p 5- qt there exists a ftrnction p that is q on paverage, where q is an increasing poIynomial,

p(z) - ij(z) > ji(z) for all 2.

Fii r 2 I. Let us define

where co = j o g(X). Obvionsly s is a poIynomial, and it f o h that

for all strings 2. Let D, = (z E E 1 p ( z ) 5 q(Iz1- 3r) 1g(x)1 5 ~~(121-3r)) . We then have

By om assumption, the first two terms are bounded above by 1/3r. For simplidty, Ilet T, represent the third

tena It &ces to show that Tr 1/3r m the rest of the proof*

Take an arbitrary z E D,, say Izl= n, and let t = g(z ) . I f f (g(z)) > s([z[ - r), then f (2) > pl(p9(3m)

6q(3m) - m2). Since Izl 5 pg(3m), we h a . j(r) > pt(lrl 6q(3m) - m2). Moreover, it holds that j(z) 5 q(3m) - ij(x). Hence, T, is estimated as fobws:

Smce ijg* (2) 5 P(r), we conclude:

In the following, we demonstrate some d the basic properties of manjr-one redua'bilities

CEIAPTER 5. AVERAGE POLYNOMLtll; TIMI3 REDUCIBILITlES 154

Proof. (1)-(2) Clear h m the definitions.

(3) We shaU show the transitivity of SF. First we amme that (AI,k) s k (A2,p2) via fl and

(A?, CI?) (A3, p3) via f2- The domination conditions for fi and f2 ensute the eciaence of two semi-

distributions % and % such that p1 SaVP %, M Saw m, h 2 141 (f;-' (z)), and ir3 2 k.ij2(K1 (2)).

Since f~ and f2 are mafly+ne reductions, we define f as f (z) = f2 0 f~(x). Then function f reduces Al

to At. In the rest of the p m c we s h d check the domination condition for f.

h pl 5 P Q, it follows that there exists a function m, which is poIynomial on pi-average, satisfying

that m (x) - fjx (z) 2 f i ~ (z) for all z- there is another function p2 that is polynomial on p~average

such that pz(x) -%(z) 2 fi2(x) for all x. We assume without loss of generality that m(z) 2 1 and ~ ( z ) > 1

for aIl x.

Let as d e h e p(x) = pz( fI (x)) gt(s) for alI x. Notice that p is polynomial on pl-average. To see this, we first note that pz o f~ is po1ynomial on paverage by Lemma 52.4 because k.l fl(z)/ is polynomial on

pl-average and pz is po lynod on waverage. Since is also polynomial on pi-average, p ttlrns out to be

polynomial on pl-average. Set tj(z) = jh(z)/p(z) for an z. Cleariy 7 is a semi-distriiution.

For any string y in ran(f), letting D, = G'(~) ft ran(fi), i t foIlows that

For each string w in ran(f~),

and thus, we obtain that tj(w) 2 tj(C1(w)). Hence,

Theorem 5.2.6 Let C E {P, NP, RP, BPP, PSPACE). The waage compCaiCy c h Aver(C, *) is dosed

domnmard anda I~-redodions.

Roof. We begin with the dosnre property of Aver(P, *) under S~-reductions. Let us assume that

(Ap) (By v) via f and (B, v) E A-, *). Note that f is computabIe m time poiynomid on p

merage. Let M be a dderministc 'Ihring rnarbine which wrnputes B in polynomiat time on waverage

We can construct another deterministic Thing &e N which, on input x, computes f (z) k t and

then simdates M on input f (x). This madhe N actually computes A, and its running time on input x

needs at most

c - (Tmef (4 + T = e ~ ( f (2)) + 11, whm T i i ( ~ ) denotes the time reqaired to compute the d u e f (z), and c denotes some appropriate

constant independent of z.

Notice that k . i f (=)I is polynomial on paverage. Since Xz..TiieM(x) is polynomial on v - a v e , the

domination condition for f implies, by Lemma 59.4, that kT"eM(f (z)) is also po1ynomial on p-average

Thdo re , N is polynomial-we bounded on payerage.

The proofs for the dasses C E {NP, RP, BPP, =PACE) are simikr. n

5.2.2 Polynomial Time Isomorphism

A ~~~~time I s ~ l z o r p ~ m worst-case complexity theory is a P-cornputable, pmvertible bijection

which pm-reduces a set A to another set B. E d an isomorphism exists, then we say that A is pisomorpiric

to B. Interestingly, most NPamp1ete sets are known to be pisamorphic.

Wang and Beknger [I121 introduced a polynomial-time isomorphism between two distributional problems.

This subsection win ~troduce the notion of polynomial-time ~ o m o ~ z ) ~

We first introdurn a notion of one-one redua%ility.

Defhitio~ 5.2.7 ( P o ~ y n o ~ T 1 ~ 1 e One-One Reductions) For two distriiutional decision problems

(A, p) and (B, v), the problem (A, p) is p o 2 ~ - t i m e one-one d d h @I-dnable, for short) to (B, v),

denoted by (A, p) 5; (B, v), if there exists a oneone, P a m p d I e reduction f wbich pm-reduces (A, p)

to IB, -

Since the reduction f is oneone, the domination condition for f is simply expressed as j+ S P v, or

eqaitmldy p SP v 0 f by Lemma 3.4.10(2).

We then m.ttoduce a polynomial-time isomorphism among distributional decision problems.

De!5itiotlS.2.8 (Polynomiany Isomorphic) [I121 For two distrriutiod decision probiems (Ap)

and (B, u), (A, p) is pI#p Lomorphic ( p io rph i c , for short) to (B, v) if there exists a P-compntabk,

phmrti i le bijection f on F such that (A+) 5; (B, u) via f and (B, u) $ (A, p) via f -'. This f is d e d

a pol@nnd-time isomorplrism (piirnorphism, for short).

Berman and Hartma& [I01 sbon by an aadokg of MyhiIl's isomorphism theorem m recnrsion theorg

that, fix two frmctions which are one-one, hmgth-irncreasing, P-campntable, and pgiatrertIe, if A 5; B via

f and B 5; A via g, then A and B are pisomarphic In fkt, howwer, we need ody the condition that fog

and g 0 f are I m Q e a S l n g instead of both f and g being Iength-imcreaSing. Wang and Belaager [lU]

give an analogous result in awqpczse settirig.

Proof. ,4ssume that f, g, p, and v satisfy the conditions of the proposition. Let p be a polynomial

which is a timebound of P-mmpntabk functions, f, g, f -I, and g-'. We 6n-t dehe two sets Rl and R2

a5 f o b

where (g o n k ( z ) means k applications of the function g o f to z, and in particular, (g o no@) = z.

We claim that Rl U R2 = 3' and Rr fl R2 = 0.

Proof of C k Assume that there is an element x in Rl n R2. We take two strings y and z such that

z = @ o nk[ar) = g o ( f o g)"(z)+ y 6 g ( C ) , and z # f (Em) for some k ,m 1 0. Notice that z E g ( C ) -

O b v i d y k > 0, since o t h m k , y = z and z 4 g(EW), a contradiction, Hence, we have k > 0. H the case wherem 3 k , we have z = g o ( f og)m(z) = ( g ~ f ) ~ o g ( z ) . -4s z = (gof lm(y) and go f is one-one, it follows

that y = @ o Am-' o g(z) = g o (f o g)m-k(z). CIearIy the bst term belongs to g(F), so y is in g(Z9) , a

comadiction. Hence., rn < k. ~ s f n g g o ( f ~ g ) ~ ( r ) =z, wehavex= ( g o n k ( y ) = g o ( f o g ) k - f o f ( y ) , and

thns (f o g)"(z) = ( f o g)k-l [ f (a())- Thus, z = (f o g)k-m-l(f(y)) = f o (go f)k-m-f(y). This implies that

z E fp), a wntradiction. Theref' Rl n R2 = 0. I

h f o f Claim TakeanarbitrarJTzinC- A s s m n e t h a t z = ( g ~ f l ~ ( y ) a n d z = ~ o ( f o g ) ~ ( z ) h

some y, z, m, and k. We also assume that k ,m are muimal- I€ either y g' g(P) or z fZ f(!Em), then z

is in Rl U R2. Now we assnme to the contrary that y = g(wJ and z = f (to,) for some appropriate w9

and to,. h m y = g(wg), it f o b that (g o nk(y) = g 0 ( f 0 g)*(w& h m z = f (KG), it foII0w9 that g o (f o g)"(z) = (g o f)m+l(m,). Botk impIy the same x, and thus we have (g o n L ( y ) = (g o f)"+'(w=)

andgo(fog)k(m,) = g ~ ( f o g ) ~ ( z ) . B y t h e ~ d k a n d m , weco~dndeboth that k = m + l and k = m Thisisacontraclictioa I

It is possible to check whether r E Rt or z E R2 within O((1zI + 2}P([zi)) steps- This is shown as

CaAPTER 5. AVERAGE POLYNOMIAL TIME REDUCIBfllTIES 157

foliows. Notice that f - I o g-l is Parnputable because ran(f) and ran(f) are both P-computable, aud,

more important, it is lengthdecreasing (ia, lzl > I f - ' o g"1) because g o f is length-increasing. Let us

consider the strictly descending cha i~

und f - I o g-' can no longer be applied. This chain comkts of at mast Jzl+ 1 elements. Let F be the last

element of this descending chain. Hence, z = (go nk(F) for some k 1 0. Notice that if F E RI , then all

other elements are in Rt by the d M o n of Rl. In particuIar, z E Rl if and only if t' E Rl. To rhedr

whether z E Rl, we w d d check whether g-' cannot be applied to z, and thus we need O((lz[ + 2)p([zl))

steps. Similarly, we can determine whether z E R2 in O((/z[ + 2)p(lzl)) steps. As a consequence, Rl and R2

are both P-computabIee

SimiIarly, we define Sl and S? as foIIows:

We also have S1 n S2 = 0 and SI U Sz = C*. Moreover, $ and S.L are P-computable.

Let us dehe the desired pisomorphism h as foJlm:

This function h is total became R1 U R2 = C*. Also h is one-one because f and g-' are oneone, and h is

also onto b a m e h(Rl) = S2 and h(R2) = Sl. It is not hard to show that

Hence, h is total and thus a bijection on C'.

We can easily see that h reduces A to B because h(x) = f(x) for all x E RI , and h(z) = gdl(x) for all

z f R2. simn;UIy, h-I reduces B to A.

What remains is to check the domination conditions for h and h-'. It is sufticient to pnme that p 5 P vo h

andv ~ p p o h - ~ - Weshattshowthat p Sp v o h , S i p SP v o f , Iet so beapolynomiasuch

that b(z) 5 so(jzj) - P(f(z)) for all z- S i l y , since p o g S p v, there is a polynomia sl such that

b(g(z)) 5 sl(Iz1) C(z) fOf all Z. Recall that g is phonest, and thus we can take a polynomial q such that

121 5 d1dz)1) far an 2- Let s(n) = +st(q(n))- Ifz E RI, then b(z) 5 ~ ( l z l ) - b ( f ( z ) ) I s(lf l)-fi(h(z))- I f z E R2, then Iet z = g(w), so we have b(g(w)) I so(lw1) f(w) I so(q(b(w) 1 ) ) - fi(ut). This implies that

f (t) I s(]z[) f(h(z)). Therefore, p 5 p o h S i y , we can show that v s p p o h-'.

Tbis completes the p d n

The reader who is iderested in other types of many-one reduciiiIities may refer to [36,9,30,51,44,32,

108)-

C&APTER 5. AVERAGE POLI'NOMIAL TlME REDUCIBlllTIES 158

5.2.3 Deterministic Ttlring Reduu'bility

We turn our attention to Taring reduc i i i . In worst-case comptexity thwy, Cook [22] first formulated

deterministic polynomial-time lhring reductions to show tbat SAT, the satisfiab'ity problem, is one of the

hardest problems that is in NP. A Thing reduction h m a set A to another set B is a deterministic oracle

Turing machine that computes set A with the help of set B as an orade.

In a- co rnp lw theory, Ben-David, Chor, Goldreich, and Luby 191 introduced a similar notion

ofdetermiaistic 'Ilrring reduciiility among distniutional decision problems. We begin with this deterministic

flrring r e d u c i i i . Let us recall that the notation Q(M, B,z ) denotes the set of strings queried by an oracle

'king marhine M with oracle B on input z.

Dehition 5.2.10 @etr ' ' tic Turing Reductions) [9] Let (A, p) and (B, v) be distributional

decision problems.

1. (A, p) is polynomial-tinu M g dud& (pT-reducibIe, hr short) to (B, v), denoted by (A, p) 5; (B, v), if there exist a deterministic oracle 'Ihring machine M and a semi-distnation 7 such that

(i) (EEaencg) M with oracle B is polynomiaI-time bounded;

(ii) (Validity) A = L(M, B); and

(iii) (Domination) C( SP q and 3 2 Xz.ij({z [ z E Q(M, B, 2))) .

2. (A, p) is atmnge polynmniol-time reducible (avpT-redtfdble, for short) to (B, v), symbolidy

(A,p) 9 (B, u), if there exist a dderministic oracle Tmhg machine M and a semi-distribution q

such that

(i) waency) M with oracle B is po1ynorniaI-time bounded on paverage;

(ii) (Valim) -4 = L(M, B); and

(ii) (Domination) p s q q and f 2 Xtj({z I z E Q(M, B,z ) } ) .

The condition (iii) is d e d the dom%&m condition fw the ndrrdion M. originany Ben-David et at [9] used a stronger domination conditioc

for some polynomial p.

ProoE (1)-(3) Clear from the debitions.

(4) Here we show that is transltrve. - . The proofs for the transitivity of the other redncibiIities are

analogous. Now we assume that (D1,pl) is avp-T-redua31e to (D2,p2) via a reduction MI and a semi-

distribution y, and also assume that ( 0 2 , ~ ~ ) is avpT-reducible to (D3,p3) via a reduction Mz and a . . .wm&miiution y. We assume that Q(M1, Dz, z) # 0 for infinitely many z since, otherwise, (D1,pl) E

Aver@', *) and, thus triviaIly (Dl, PI) is avp-T-reducible to (D3, p3). In what follows, we shan show that

(DlIcrl) (D3tP3). By dehition, there exist two functions fl and f2 which are poIynomiaI on PI-average and on praverage,

respeaively, such that fr (z) - 4 (2) 1 fit (x) and f2(z) - 4(z) 3 F2(z) for an 2. Wdout loss of gmerality,

we assmne that f~ (z ) 1 2 and f2(4 1 2 for aI1 strings z.

We dehe a new machine M as f o l l m on input z, M deterministidy simrtlates MI on z, and whenever

Ml queries a string y, M detefmipistically simulates Mz on input 3. In the case where s is the empty string

A, M must be designed not to query any strings (if M2 queries some strings to o d e D3, then their o&

answers are encoded into a program of M). Clearly Dl = L(M, 4 ) ) . Note that

Next we shall show that M and f are both on pr-average We 6xst see that M is @pomiaL

time h d e d on PI-- Let h(z) = b ( ~ , ~ ~ ~ ~rmez(y) fix each z. To compIete the p& by

CHAPTER 5. AVERAGE POLYNOMIAL 2'1343% REDUC3BaJTIES 160

Lemmas 33.13 and 3.3.12, it is safficieut to show that h is polynomial on pl-average since k . ~ i r n e Z (z) is

polynomial on pl-itverage; however, this is not difficult to see.

The proof that f is polynomial on pl-average is sbdar, and thus the claim is established- 0

The transitivity of avp-T-redua'b'i implies the dosure property of A m p , *) under am-reductions.

Proof. Assume that (A, p) (B, v) for some (8, v) E Averp, *). Note that

Sice (B, v) IF (0, v-), by Proposition 5%11(4), we obtain (A, p) s! (0, vapad). Thus, (A, p) E

A=(P,a. 0

5.3 Many-One Complete Problems

We have introduced two types of reduaiilities in the previous section which play a significant roIe

in measming the relative compkity of any two distributional decision problems. If a distriiutional probIem

(A,p) is d u a i I e to another distniutiod probIem (By v), then (B, v) is considered at Ieast as hard as

(A, p) to solve. We wish to see the hardest problems m Di(NP, P-comp) in tbis sense.

The notion of compkte problems was introduced into computational compiexity theory in the early 19707s,

and subsequently many problems have been found to be complete for NP. We generalize the dehition of

"completeness" to our setting below.

Dehition 5.3.1 (Complete ProbIems) Let I, be any reduciiility and Iet C be a cIass of distributional

decision pmbIems,

1. A distnintiond probIem (D, p) is Il,-lrmd for C if every problem (E, q) in C is 5,-reduaiIe to (D, p).

2 A distributiond problem (D, p) is ~,-compiek for C if it is in C and is <,-hard for C.

This section wiII show that several impoaad distniutional decision problems are pmcomplete for

Dist(NP,P-camp), and hence that are among the hardest probIems to sohe.

Intrigningly, Behng~~ and Wang [6] pointed out that most known pm-compIete prob1ems fix Dist(NP, P-comp)

are actnally pixrmorphic. They proposed an version &a conjecture given by Berman and Hart-

manis [lo], known as (Berman-Hartmanis) isomorphism conjecture, that asserts that all pm-amnplete distri-

butional probIems for Dist(NP, P-oomp) are p-isamorphic We refer to this conjecture as the Isomorphism

C ~ - e d w e

CFUPTER 5. AVERAGE POLYNOMIAL TIME REDUCIBIllTIES 161

5.3-1 Randombed Bounded Halting problem

One of the most nsefnl distxiiutional decision probIems is the rnndomLed bounded haitkg problem To

descn i it, we assume an &ective enumeration of an nondeterministic l'hing machines, say { M i ) i E ~ - The

rmrdomized bounded hulting problem (RBHP) is the distributional problem (BHP,wm) that is based on

the bounded ttalting problem

BHP = ((%,x, 1") I Mi accepts z in less than n steps ),

where si is the string corresponding to i and based on the defauIt ptobability

Intuitively, we independently pick up a string s,, a string x, and a wary string 1" at random. Clearly F ~ H P

is supportive and P-computable. We remark here that the choice of Pay(ln) is essentd.

In the foIIowing theorem, we shaU prove that RBKP is SP,-aomplete for Dist(NP, P-comp). This theorem

has been proven in numerous ways (see eg-, [9,30,36,1111). The proof below follows the argument given

by Wang and Belanger [Ill].

53.2 [36] RBHP is<P,-c~mpIetef~rDist(NP,P~p) .

Rod We 6rst note that RBHP is in Dist(NP,Pcomp) because BHP beIongs to NP and ~ B E P is

P-computable.

Consider any distniutional decision problem (D,p) h m Aver(NP, P-comp). There exists a poIynomial-

time nondeterministic 'Ihring machine M which accepts D. For every set D E NP and every distri'bntion

p E P-eomp, we next show that (D, p) I D , Let g be the one-one, pinvertibIe, P-computable

fnnction of Lemma 4.2.7(2). Note that lg(x)l I q(Ix1) for some absolute polynomial q, and that b(z) < 2-b(=)1c2 for a11 x. Now let us consider the machine N that simulates 116 in the following fashim

begin nondderministic 'bring madhe N

inpat Y compute a string z such that y = g(x) by b i i search

(this is done in time p o l y n d in [zl)

if such an z exists then nondderministically simdate M on input z

else reject

end

S i . N is a polynomiat-time nondete * -+ic 'lSxing mnrfiinp_ we take an index i such that L(Mi) =

L(nr). Let p be a polynomial tim&ttud of Mi- The desired reduction f is now defined as f(x) = (si,g(~), ldl'l)). It is not diBeoIt to see that f is one-one and reduces D to BHP. It is d e n t to

check that f sathEes the dcmbatitnt 001~ditioa Let s be any pdynomial such that fi-(si) - ~ ( l t ) 2 256(q(n) + l)'(p(n) -t I)* for all n E N Note that i is a constant and does not depend on R Then, we have:

CHAPTER 5- AVERAGE POLYNOMLQL TIME REDUCIBILEIlES

In Section 6.5, we shdl generalize the problem RBHP in order to show that the generalized RBHP is

also complete for Di(CE, P a m p ) for ea31 k > 0.

5.3.2 Randombed Bounded Tiling problem

The rrmdomized bmmded tiling p b k m (RBTP) is the first problem discovered by Levin [60] to be Ig- complete for Dist(NP,Pamp). RBTP is the distn'bntiond problem (BTP, pm) that is de6ned as follows.

A tile is a quadruple [u, v,z, w] of strings, where u is called 'left: v "top," z ''rightP and w "bottom." We

use the notation le f t[u, v, z, w] to denote the left element u. SimiIar notations are used for "top," "right,"

and "bottom." Let Sn be the n x n square (1,. . . , n) x (1,. . - ,n). Let T be a set of tiIes. A frmction f h m

S, to T is d e d a T-tiling of Sn if le f tu( i + 1, j)] = tightCf(i,j')I and b&om~(i, j + l)] = top[f(i, j)l for

a l l i , j w i t h l S i < n a n d l 5 j < n . Aseqnence (t,, ...,&) isa T-romofIength k i f t i ETforaIIiwith

1 5 i 5 n and left[sj+l] = right[sj] for all j with 1 < j < n. Let

BTP = {(T, In, lk, (tr , . . . , tk)) I (tl, . . . ,tk) is a T-row of length k, 1 5 k 5 n,

3 f V is a T-tiling of S, and Yi[l I i 5 k + f (1,l) = t in}.

Fhc a positive P-compntable distn'bution u for a set T of tiles and let

~ ( ' Z ' ) - h y ( l n ) - f - ~ ~ l ~ i f l s k < n a n d

k ( T , In, lk, (h, - - - 7 tk)) = { 0 T i # 0 f o r a l I i w i t h l i i < _ k ,

0-

where Ti = {t E T I left[t] = right[&]). We remark here that the choice of a default distriiution w for tiles

is not important because it does not affect the proof of Theorem 533.

It is shown that RBTP is pm-redmi to RBHP [a, 361.

Tkomm 5.3.3 [I121 RBTP is pisOmOrpAic to RBHP.

Proof. It is known that BTP is NP-complete. Thns, RBTP is m Dist(NP,Pcomp). In the following, we

shall construct two ontxme, Iengt&mmaing, pberti'iIe, P-computable reductions f and g b m RBHP to RBTP and h m RBTP to RBHP, respectively, with the condition that sP ~BTP o f and pmp szP

p~m o g- This is sufficient to show the theorem, baawe l'mp&bn 5 2 3 yields the existence of a p isomorphismbetweenRBHPandRBTP.

Since BHP E NP, there is a nondeterministic 'Ihring machine M accepting BHP m polynomial time

l k m the Distrib&*on ControIling Lemnta (Lemma Uf), there E&S a total, oneone, phmtible, P-

CHtLPTER 5. AVERAGE P O L Y N O ~ TIlldE REWtTCIBIUTES 163

computabIe trmctian h snc6 that 4 - 2 - i h ( = ) [ < bHp(x) 5 20 - TI''(=)[ 6Df an input z. Tak a polynomial q

such that [h(z) 1 5 q((z1) for a11 z. For each string w, let WL = Omlsl,lw, where m = IsltPI[. Notice that if a string z is given, t h we can

uniqudy determine WL such tbat WL [5 z if one exists. Let M' be the following algorithm:

begin algorithm M' input z

if there is no w such that ut~ L z then reject

(Assume that there is the uuique w such that WL t z.)

check if w = h(z) for some z by a b i i search

(This is done in time poIynomkd in 1x1.)

ifno such x exists then reject

simnlate M on input z

Now write f instead of h(z) for brevity. We then have ZL = Omlslh(,)l h(z) , where m = I$lh(r)ll- Note that

Theftmaion f i s ~ e a n d p i r n r e r t i b 1 e ~ s o i s b Werema&that,hreachiwith 2 < i _ < k , the llllmba of tile r which nuVhes e,ui,#, q to the right Cre, leflftl= W@,lti, #,$I) is 4~ 2, and

comeqdy, the probabiiity that the ith tile of S, is chosen is 1 f2

CEW'TER 5. AVERAGE POLYNOMIAL TIME REDUCIBlU!lZB 164

P m f of Claim Assume that M' accepts astring ZL. By the choice of the initial row of length k, the sequence of the top of the tiles set in the bottom row ofthe square is exady 5~2, where r satis6es [ 5 ~ z l = ~ ( 1 ~ 1 ) . This

is an iaitial ID (ie., instantaneous dewiption) of the rnarhine Mt on input 5~z . The second raw of tiles

tbat match the bottom row repnsents the ID of M' on ZLz obtained by a single application of the transition function d (ie, after one step) because a pair of tiles in (T4) (c~tfe~ponding to J(Po, a) = (4, b, R)) are the

onIy tiles that are diffewnt ftom the symbols on the top of the bottom row. Fkursively we on continue

this argmnent, and then, de~ we rebuild the bistorp of the ID'S of Mt on input &z, we see that the tiles

fill the square. Recall that Mt reaches the accepting state m Iess than p(lz1) steps. Conversely, if the tiles

fllI the square, the sequence of the top of each row dsmi'bg an ID of M' on input ZLZ. This implies that M' accepts 5 ~ . Hence, f becomes a reduction from BEEP to BTP. I

We shodd check the domination condition for f. Let s(z) be a polynomial such that

for all z. Then, we have:

By the -on of s, the last term is bounded by:

CaAPTER 5. AVERAGE POLYNOiUtAL TIME REDUCIBlU!llB 165

Let p be a poIynomia1 which is a time bound of M'. It is obvious that z E L(M) if and only if

ht(z) E L(M3 for all strings z. Let i be an index such that L(Mi) = L(Mr). Let p be a p o l y n o d such

that TmMi (a(%)) 5 p((z() for all Z. D&e g(z) = (si, hr(z), 1P(Iz 1)). Clearly g is one-one, length-inQeaSing,

P-computable, and pinvertiile. Assume that Ihl(z)[ I q(lz1) for some poIynomia1 q. Take a polynomid s'

so that

~'(2) 2 1280. @(z) + 1)2 ( q ( ~ ) + 1)2/V-d(~i).

The foIlowing catculation is similar to the one above.

Hence, we have crrrrp 5 P pBm o g. It is also easy to show that cz~ap o g AP CCBT~. Therefore, ~ B C ~ B E I ~ o g =P

p~m. This completes the proof. n

Knijnenburg 1511 considered complete problems for Dist(PSPACE,P-comp) and pointed out that an

extension of randomized bounded tiling problem becomes <P,-compIete for Dist(PSPACE,P-comp).

5-3.3 Other Complete Problems

BPCP = {((L), 1") I L = ( ( U I , ~ ) , -. . , (%,urn)) A 3k 5 n[ there exists a solution of length k for L 11,

where {L) is the encoding ((ul, q ) , . . . , (%,tt,)), and the distn'bntion ~ B ~ C P is defined by

This deFanIt probabilitp is experimentally given by picking np independently and randondy tm natnral

ntlmbeR nand m, and 2m stxiqs U ~ , - - - , ~ ~ U I , - . . ~ V , , , -

This problem RBEIP is gm-redncible to the randomized bounded Post correspmdence problem RBPCP.

Another problem known to be complete is the Rhndmnized Pdhfmme PmbIem, which was defined by

Gurevich [MI. W q and Belanger [I121 intraduced the RundOmited W d P d h for huc System and

showed that this probIem is pm-compIete for Dist(NP,P-comp).

5.3.4 Hard Problems under Samplable Distriiions

We have seen several intrigning complete problem for Dist(NP, P-comp). A natmal question is: does

the dass Distw, P-samp) have any p-m-eomplete problems ? Ben-David? Chor, Goldreich, and Luby [9]

constructed such a distrr'btrtional problem-

F i of all, we shall dehe a "universal" P-samplable distriiution pu- Take arr e f f d v e enumeration

{%)iEn of an O(n)-time samplabie distrritttions. Let us define pu as follows:

Experimentany, we choose a number i at random, then sample a string under the distribution ~ i . Clearly

pu is a P-sampIabIe distrriution.

Proof. WE must reduce an a r b i i distnitrtiod problem (D, p) in Aver(NP, P-samp) to (BHP, pu).

By Lemma 643, we can take a strictly P-samplable distriiution v such that p i p v. By its samphbilitp,

there exists a randomized lhring machine M sampling v in time q in the length of its output, where q is

an appropriate polynomi=rI, For each z, let g(z) be 20q(1'1)-lrl* Let us define P' by up. W e show that

this distrr'bution J becomes O(n)-sampkble To sample d , we consider the foIIOwing procedure M': on

inpat A, simulate M, and if z is an output of M, then output z~~(w)-l+l* We fhr any random seed s which

leads to the output z. The rrmning time TimeM~ (X; s) is O(T'ihne~(X; s) + q(W)) E O(q(1zl)) because

TimeM(& s) 5 q(lz1). As a dt, it is bounded above by O(n) m the kngth n of output-

We then dehe the set D' by

06vi0tlsly7 lY is in NP since D is so. By the definftion of 9, it fonows that z E D if and only if g(z) E D',

a d - (D? P) I p , (0, t/) 9-

Now we must show that (W,J) is pm-redua'b1e to (BHP, pU). Because D is an NP set, it is recogniz-

able by some nondeterministic 'Ihring machine, say M, m po1ynomid time Let p be a strictly incfeasing

poIynomiaI which bounds the nmuhg time of M. Let i be an index so that L(Mi) = L(M) m the list

{M')im- We take the standard reduction fmtctim f diseassed &re: f (x) = (e ,x , Wl)} for dl z.

Let Timef (2) be the amputation time needed to compute f (z) determiaistldy. Smce f is P-compntable,

for some constaatg d > 0, Timef (z) is bounded by

CHAPTER 5. AVERAGE POLYN0MU.t TIME REI)UCIBIllTIES 167

Thns, f is computable m O(n) steps in the Iength n of its output.

Now define $(z) = g(f-'(z)) for aII z. This q is O(n)-time samplable, and thus there exists an mdex i

such that q = Q. By the definition of p t ~ , fi&) 3 2--(* I-' - iji(z). Then, we have:

Gf(2) . i u ( 4 2 - - 8(i + I)*

Therefore, (A, p) is pm-redua'ble to (BEP, pu) .

5.3.5 Discussion of Complete Problems for Aver(NP, P-comp)

W e have seen several complete probiems for Dist(NP,P-comp) and Di(NP,P-samp). In this subsection,

we shall discnss complete problems for the average-case complexity class Aver(NP, P-comp).

Udortnnately, it is not known whether Aver(NP,P-comp) has any SP,-compIete problems. This is

because of OW d M o n of nondeterministic 'bring and their accepting criteria In this section,

we shall see what happens when we take other mod& of nondeterministic ' k ing machines and accepting

aiteria

However, by W a g and Bdanger [Ill], if we use a 9clo&edn model of nondeterministic lking machines

(thus, all computation paths are assuxned to be of the same length), then any pm-complete pmbIem for

Dist(NP,P-mmp) is complete for Aver(I+W,P-mmp). RecalI the notation Aver(NPb,F) used in

Section 3.6. We shall see a complete problem for Aver(NP',P-comp).

We begin with a general result. Let as expand a notion of "weak Cdescriptiven to u~adeScriptfve" using

any 5,-reduuiiIity.

Dehition 5.5.5 (5,-Desctiptive) Let 5, be any reduaiility. An averagecomplexity cIass Aver(V, a is <,desm+ptiae if, for every problem (D, p) in Aver(V, F), there exisfs a probIem (E, v) in Dist(V, F ) such

that (D,P) So (E,v)-

ProoE Assmne that (A,p) is Sz-complete for Dist(C,F). Let (B,vj be an arbitrary distnintional

problem m .4ver(C, 7). We shaII show that (B, v) is 5,-rednuile to (A, v). Smce Aver(C,F) is 5,- descriptive, there is a problem (B',r/) E Dist(C,F) such that (B,v) is 5,-rednuile to (B', J). From the

asmmption that (&p) is Ipm-oomplete for Dist(C,F), it foIIows that (B',J) 5, (A,p). The transitivity of

s a implies that (B, 4 I* (4 PI* 0

CaAPTER 5. AVERAGE POLYN0MU.L TIME REDUClBILITlES 168

Proof of Cluin~ It d c e s to show that AvercNp',P-comp) is I~desaiptiveve Assnme that (A,p) E Aver(NPD,P-comp). Let M be a nondeterministic M g m;uhinP which computes A m time polynomia

on paverage- Let as take g as in Lemma 42.7(2). Note that g is P-compntable and pinvertt'ble. Then, we

have $(z) < 2-L9(f)1+2 for aIl strings z. We define v as

Ym@(z) ) &mY(ln) if z = {z, In), B(z) = otherwise.

It is not difficult to see that u is a P-computable distriiation and does not depend on p.

To get the desired result, we wodd define a --one reduction f and a set B in NP. Let f (z) =

(g(z) , lrme~(*)) for every 2. We claim that (f, p) E Aver(FP, P-comp) . Note that the fnnction Xz.Time~ (z)

is computable in time polynomial on paverage. Since g E FP, f (z) is computable in time po1ynomial on

paverage. Note that f is one-one. Let B = f (A) , It foUows that 3 E NP since B is computed by the

foUowing lhring machine M':

begin nondeterministic agwithm for M'

inpat b , l n ) compute u f a which f (u) = z

simuIate M on input u for time n

if M fails to halt then reject

output M ( z ) and halt

end.

This implies that (B, u) E Dist(NP, P-comp) . Now we set q(z) = 256()g(x)I + time^ (x) + 1)'. Then,

by Lemma 33.14, q ttrms oat to be pdynoiniaI on paverage.

We next show that &(f (z)) q(x) 2 fi(2). For all 2, we have

If we use the length ofthe longest cornpatation path whenever the madim rejects an input as the rrmning

time, then any pm-complete pmbIem for Dist(NP,P-comp) is a;ppT-complete for Averw,P-comp).

C h h 11 RBHP andRBTP am rrtrpT-complete forAver(BW,Pcomp), arlrereAver@W,Pcomp) is

d e f i n e d ~ o n t h e ~ o f n o n d e t s m a r t s t s c . - - ~ ~ 1 D i t h r r r n r r i n g t i m e d b p f h e ~ e s t

n$xtingpath deneuerit ~ * & cm irqmt.

CHAPTER 5. AVERAGE POLYNOMIAI, TIME REDUCIBaJTlES 169

Pnmf of Ctaim For this daim, it suffices to show that Am(NP',P-comp) is ~Faescriptive. Assume that (A,p) E Aver(NP0,P-amp). Let M be a nondeterministic Turing machine which computes A in time

polynomial on paverage. Assume that g is defined as in Lemma 457(2). Note that g is P-computabIe and

pmvertiiIe. Then, we have fi(z) < 2-Mf)1+2 for aI1 strings z.

RecaIl the proof of CIaim 3 in Section 3.6, and consider the same set C and the deterministic o d e

'Ihring tuachine N that computes D with oracIe C m polynomial time on paverage. Remember that, for

each query string (b, z, ln), z is the only input string that N on input z can query (b, x, 1") to oracle C. Let

$ fimd(x) - ;ully (1") if z = (b, 2, In), P(z) =

otherwise.

It is not difficult to see that v is a P-computable distribution and does not depend on p or C.

We modify the 'Ihring machine N into Nt m the foUowing fashion:

begin algorithm Nt

inpat z

compute u for which g(u) = z simulate N with oracle C on input u

output N'(z) and halt end.

This implies that (B,v) E Dist(NP,P-comp). Now we set q(x) = 512(1g(z)I+ 1 ) 2 ( ~ i e $ ( x ) + 1)'. Then

by Lemma 32-14, q tnrns out to be polynomial on paverag~

For all z, we have the simple estimation, where n = ~rme('N,(x),

5.4 hcompleteness Results

W e shall discus two important distriiutio= fiat distributions and sparse distriitrtions, both of which

possl'by make amdated distriiutid problems incomplete for Dist(NP,Pamp).

54.1 Flat Distributions

CZAPlEX 5. AVERAGE POLYNOMIAL TIME R E D U C l B m 170

We W obserrre the distn'butions of some pm-complete problems discussed in the previous sectio~

Neither p~m, m? nor mpcp, fDr example, is a Bat distributio~ Here a distribution p is fiat if there

exis& a red number E > 0 such that ji(z) 1 2-lrV for all z. This is seen as follows. For example, assuming

that p ~ a ~ is flat (ie., ~~BHP (z) I 2-br for some t > 0), let us consider sufEaently Iarge n, i, and z satisfying

n > (Uog(i) f [ z [ )~ /~ . Then we have

On the other haad, we have

Hence, using the assurctptiion that map is flat,

2--i)-l*l Now we have suaecP < T ~ * ~ - ~ ~ ( . ' - ~ I * ~ , and t k pOg(*lrl*a" < n3 f l ~ ~ ( i ) ~ - /zI2. Thb is &&y

a contradiction. Hence, is not flat.

One might suspect that there is no <P,-complete probIem with a Bat distribution. Gurevich [36j, and

Wang and Belanger [I121 indeed showed that distributional problems with flat distr i iutio~~~ are not complete

hr Dist(NP,Pcomp).

Using the same argument as above, we can show the fonowing result. Under phonest, P-computable,

one-one reductions, there are no complete problems with flat distributions in Dist(NP,Pcomp).

Proofl We s l d prove the conhpodim Assmne that (A,p) is <~-compIete for Dist(NP,Pamp)

under omone, phones~ reductions As RBHP is in Dist(NP,Pamp), there is a phonest, Pampatable,

oneone reduction f that minces RBBP to (A,p). The domination condition for f, by Lemma 3 - U ( 1 ) ,

+hesthat ~BEF SPpof since f isoneone. Since-ismtaflatdistn-brrtion,thenpof isalsono~flat-

CHAPTER 5. AVERAGE P O L Y N O ~ ~ TIME REDUCIB~LJTTES in

As f is phonest, p onnot be flat. To see this, assume that @(x) 5 2-IZr and If (z)lk 5 1x1 for r, k > 0. Thus, Ff (z) 5 2-If (41' 5 2-14? 0

Pmpodion 54.2 [S6] A m m e that EXP # NEXP. For any pat distribution p and any EXP set D, the d 5 t r k h d problem (D, p) is not aopm-hmd fm Dist (NP, P a m p ) .

Roof. Assume that (D,p) is 5?-hard for Dist(NP, P-comp) and show that any set in NEXP is deterministidy solvable in exponential time

Let A be an arbitrary set in NEXP. There is a nondeterministic Thing marhine M which computes A

in time 2p(Ifl), where p is a 6ked polynomial We may assume that p(n) > n + 2Uog(n) + 1 for all n E N For each x, set z! = zOl""*')-l'l-'. We note that if 1x1 = n, then Id) = 2dn). We set At = (z' I z E A).

This At belongs to NP and consequently it is in EXP. Let MN be a d e t ~ ' -stic 'ILring machine which

compaes A' in exponential time.

We define J as follows: .

To see that J is PamputabIe, notice that u'(z)

lim;, V(Z) = &- v-d(u) = 1. Smce (A', J) E Dist(NP,P-comp), om assumption impIies that (A', J) is avprn-redua%le to (D, p)-

Take a reduction f witnessing (A', d) IF (Dl& Note that f is computable in time polynomial on J-

average, and J 4.- q and ji 2 I jy for some semi-distriintion q. We note that, by the dehition of J, f is

actualIy computable m exponential time.

Fbst we show that If(z')l is bounded by a po1ynomial in 1x1. The domination condition yields the

m c e of a function s being polynomial on J a r a y such that jib) > &-l(rl for dl y. In tt 2') particuiar, we have @(f(d)) > h. h u m e that s is q on J-average for some polynomial q. Then, for

almost an x,

for ahcast alI 2. Smce p is flat, for some constant m > 0, we have

This yietds the desired that 1 f (a!)[ 5 (k + 1)" p ( I ~ l ) ~ . Let us consider the following deterministic algorithm M which computes -4.

be* det ' ' ,tic aIgorithm M input z (say, n = 121) compute 2' (= ZOI~"'-"")

compute f (4 (*I simulate on input f (d) and halt

end.

The running time of line (*) is at most s;ponential in Izl since If (d)l is bounded by a polynomial in lzl.

Hence, the total running time of algorithm M is e x p o n d in the length of input. This implies that A is

in EXP. 0

5.4.2 Sparse Distri'butions

Gurevich [36] called a distributional problem (D,p) "sparsen if the set {z I ji(x) > 0) is sparse, where a set

S is (plynomially) spm~c if there is a polynomial p such that IlSn PI1 5 p(n) for aU n E N In this section,

we instead call the distriiution p (pdynomiaUg) spurse if the set {z I ji(z) > 0) is sparse. In other words,

p is positive only on a sparse set. For instance, the standard dktriitxtion on (0)' is sparse.

De0nition 5.4.3 (Sparse Distributions) .A distribution p is called (plynomially) sparse if the set

(2 I b ( 4 > 0) is sparse.

We shall see that any distn'butionaI problems having a flat distriiution cannot be pisomorphic to the

standard compIete problem RBBP.

Proof. Assume that (D, p) is pisomorphic to RBHP via a bijection f. The domination condition by

Lemma 3.412(1) implies that there exists a pbormded hmction p such that p(x) - fi(f (2)) 2 hHP(z). For

the sake of convenience, we set S = {(si,x, In) 1 i,n E N, x E C').

Srncec(~~ ispmitiveonS,thedishibntimpo f isaIsopositiveonS;that is,

By the sparseness of the set (z I F(z) > 01, the set f (a is also sparse. In particular, f (BHP) is sparse.

Therefore, since f is one-one, BHP shodd be sparse This contradicts the fact that BHP is not sparse. D

It is known that, under the isomorphism conjecture, no sparse sets can be NP-complete- In 1988,

Mahaney [66] proved that, without assmning the isotnorp6ism conjecture, there is no sparse NP-compIde

set unless P = NP- Notice that the hmorphb conjecture mnficts with the assumption P # NP. A

similar result also hoIds in our a- environment.

Thecmm 5.4.5 A~swnt that P # NP. For ony set D and ang sparse distribuk'on p, the disCribuEioncl

problem (D, p) is not pm-hard fw Dist(NP, Pcomp).

Proof. Assmning that (D,p) is pm-hard for Dist(NP,P-comp) and p is sparse, we shall show that

NP collapses to P. In wcuIar, (BHP, mm) I L (D, p) via some reduction f, By an argument similar

to that in the proof of Lemma 5.4.4, the domination condition by Lemma 3.4.12(1) implies that there

mists a pbounded function p such that ji(z) 2 &-z(r) for all zn Thus, in particnlar, it holds that

p(z)-fi(f (x ) ) 2 ~ H P ( X ) . As in Lemma5.4.4, the set f (S) is sparse, where S = ((st, x, 1") I i, n E y x E X*). We shall use the Eact that BEP is "seIf-ted~ciite,~ that is, to determine whether (si,z, ln) E BHP, we

can check whether (s,, r, In-') E BHP or {sj1 ,z, I"") E BHP, where jo is an index of the machine that

deterministidy chooses the k t nondeterministic branch and then SimuIates Mi, and jl is similarly defined

by choosing 1 instead of 0. We view tbis process as a tree. S i we can determine whether ( s j , ~ , 1") E BHP within n steps, the height of the tree is n- Now let us consider such a self-reduction tree. This tree may contain

exponentially-many nodes, but when they are mapped by f, there are at most polynomially-many distinct

d u e s taken by f. Hence, many nodes of the tree merge. This motivates us to construct a polynomial-time

dgorithm which cornpates BHP m the b l h b g fashion.

Let us k t describe the main body of the algorithm M helm

begin deterministic algorithm M for BHP

input (si, z, 1")

setVis i t=@andsetDead=0

caIi Murkng((si, r, In), Visit, D a d )

*ect and halt

end.

Marking(tu, Visit, Dead) is the foIIowing subroutine that recnrsiveIy kills nodes which do not lead to an

accepting consgrnratim in a depth-- search:

subrouthe Mofking(zo, Visit, Dead)

if w is a Ieaf and true then accept and halt

set visit*= Visitu {w)

compute f (4 for all nodes u E Vist until neither Visit nor Dead changes

iff (w) = f (u} and u E Dead then set Dead r= Dead u {ur)

ifboth chiIiireu~o,u~ ofu are in Dead then D d .= D&u {u) and Visit -= Visit - {w, UI)

i f u is aleafand f a k thenDead F DeadU{u)

end-f~

construct a left child of w, say tug

call Markkg(wo, Visit, Decrd)

construct a right c ' d of w, say wl

calI Markkg(m, Visit, Deud)

retarn.

The aIgorithm M requires at most poIynoxniaUy-bounded nmning time and also computes BHP. There-

fore, BHP belongs to P. The andusion P = NP follows immediately from the fsct that BHP is pm- complete for NP. D

5.4.3 Unreasonable Distributions

h 1996, Paven and Selman [83j presented another type of incompleteness remlt. They d e d a distribution

p m d k if h . j i ( Z a ) E Q(n-') for some number k > 0. Any distriiutionai decision problem with an

w m s o d l e distribution fails to be hard for Dist(NP, P-comp) tmless NP is smatl (i-e, NP has pmeasure

0). Smce many researchers believe that NP is not small, this result shows another limitation of distrt'btrtions

m distriiutiond complete problem.

Thecmm 54.6 [a] Assume thut NP has pmcarme I. Let D be a set and kt p be such that ~n.ji(.C*) E

f'l(n-*) fm any positive integer k. Tha, the d i s t d t d i d poblem (D ,p ) i s not pm-had fm Dist(NP, Pamp) .

Proof. Assume that k b ( Z a ) E R(n-') for any pasitne integer k. We aLPo apmme that (D ,p ) is

prn-bard for Dist(NP, Pamp). In particular, for any NP set A, the distri'butiond problem (A, v - ~ ) is

pmmduuiie to (D, p). Smce D f EXP, there exist two posititre integers K and c such that D M o q g

to DTIME(c - 2nk* + c). For brevity, we can assume K = k. Note that this assmnptiion does not essentiaIIy

a f k t the following argmnent,

Let {fi)i*c be an effective eumneration of all poIynOmial-time comptttab1e functions- Take any set A in

NP and assume that (A,umd) Ip, (D,p) via f'. We daim the foIlOwing.

Pmof of Claim Assume to the amtray that there exist positive integers k and such that If,(z) ik > lz[ for dl Stings 2 of k t g t h 2 no. We shaIL show that satidies kz . f i@>) E Q(nd) for some d > 0. This

d d y leads to a corrtradi&cm-

implies that i(P) 2 5.nd(C1L)/q(n) for any integer n.

for any number n 2 32. Fix R a r b ' i y but greater than I # ~ . Then,

RecaU that D E DTIME(c - 2nh + c). We d e b the I-MS d as foIIows:

( 1 i f z = &

if Ib(sl=l)lL 5 Isl=ll a d b # V'(slzl) E Dl, whae b E {0, 1 ) and z = wb. It is easy to check that d is P-eomputabIe, Let us check if d succeeds on

D. By the d-on of d, &(D[O.s]) = 2&(D[O-n - I]), where &(w) = d(m, w). Thas, we obtain

lim supn- & (D[O.n]) = 00, and consequently d succeeds on D.

For each i E N, we define the set to be the set on which 4 succeeds- Notice that the eoflection {&IiEN is a punion of all NP sets. We thus condude by L e m d - 7 3 that NP has p m m 0. This contradicts our assumption. 0

5.5 Bounded-Error Probabilistic Reducibility

As seen m Section 5.4, deterministic many-one reductions are so restricted M, under the common belief

that EXP # NEXP, no distii'bntional problems wi th flat distrr%tttio~~~ are complete for Dist(NP, P-comp) . Many "natnral" distri'bnti011~ forgrapbrelated problems are rrctnany flat. Is there any hope that we can prove

such problems to be '%ad" to compute ? Vealrawsan and Levin [10q presented a solution by introducing

another redtta'bilitqr, d e d "random mauy-one reduciiEty," to m e a i m the compiexity of distrtiutional

prob1ems1 Their notion of "random many-one reducibility" was fbrther studied by Impagkm and Levin

[q and extended by Blags and Gtlrevich [I4 (see Section 53).

b m a di&mt point d view, Ben-Dad, Cha, Goldreich, and Lnby [9] introduced a new notion of

"random truth-table (and Thing) reha%*-'' This section wiI l hI1m the idea of b D a v i d et al- and introduce p r o w trrrtlGtabte ndw&By among di9tn'b~onal decision pmbIems. (This name seems

more appropriate than "random reducibilitJP because of its G m h i t y to worst-case prob&Sstic Thing

redactim)

5.5.1 Skew Bounded-Error Probabilistic Reduuiility

Ben-David, Chor, Goldreich, and Luby [9] introduced a notion of nmdom reducibility for which reduction

machines are probabilistic rather than . . . Here we use the phrase skaa bounded-- p r o m

mdh-toble reducibility to describe this type of reductions.

Dehition 5.5.1 (!&w Boded-Error ProbabiIistic 'lhth-lhble Reductions) [9] Let (A, p)

and (B, v) be any distributional decision problems.

1- (Ap) is skew bmmde&ezrw prohtbdi& pdynonicrl-time truth-table reducible (skew bpptt-reducible,

for short) to (B, v) if there exist a randomized oracle %ring machine M, a positive real number c, and

a semi-distribution q such that

(i) (Query Type) M m a k noaadaptive queries;

(ii) (EEciency) M wifh o&e B is polynomial-time bounded;

( i ) (Validity) P r M [ ~ B ( ~ ) = A(z)] 2 3 + r for all z; and

(iv) @ominattion) p s;PY, q and i 2 Xs.ri[{(z, s) I z E Q(M, B, z, s))).

2. (A,p) is skew bounded-error prob&idic ooemge poigmmid-time truth-table reducible (skew avbpp

tt-reducible, for short) to (B, v) if there exist a randomized oracle Thing machine M, a positive real

number r, and a sernidistriiution g such that

(i) (Query Type) M makes nonadaptive queries;

(in) (EEciency) M with oracle B is polynomia-time bounded on paverage;

(iii) (Validity) PrM [AIB (2) = A(x)] > $ + E for all z; and

(iv) Pominati4 p q and fi 2 W ( i ( z , s) I E Q(M, B, x,~))).

Condition (iii) is also caUed the &mhdh condiCion for the redDCti4R M. Ben-David, Chor, Goldreich, and Luby [9] instead used the foUotRing domination condition:

fm some poIynomial p Our condition (ii) is obviously weaker-

In worst-case complexity theory, we oRen assume that an orade 'k ing machine queries ody strings

whose length is larger than that of the input. The motivation is that we can always construct another d e

machine which satisfies this condition without c h @ g the amp- of computations, and also we can itmi

another oracIe set which is very dose to the o c i g i ~ I oracIe set- For example, assmne that poIynomiaI-time

oracle Turing xuachine M with o d e s e t B compdes aset A. We let B' = {(z,z) I z E B) and let M'

sixdate M with the fobwing change to orade ~nedes: if M queries a string z, then M' ~neries a pair

(z,z). Then,deady A i s ~ a m p u t e d b y M ' w i t h d e B ' d a n d B ' 9 ~ B holds W e ~ o b t a i n a r e s n l t ~ t 0 c w e m ~ ~ ~ t h e o r y ; ~ , i t s p d i s m o r e i n m , 1 d

Lemma 5.5.2 A~sume that (44 is skew aobpp#-tedudk to (B, v) with B # C'. There Qists a bvmukd-

error probbdistic om& moddne M' ctnd a d & d d h d problem (B', d ) such that

(iii) all strinqJ queried by M' tDifh o d e B' on input z ore of length greota than 121.

Proof. Let us assume that (A ,p) is skew avbpptt-reduuiIe to (B, v) via a bounded-error probabilistic

oracle Toring marhine M with nonadaptive qumMes which ntns in time p on 1-averageT where p is an inmasing

polynomial. For simpticity, Iet I' be the random-input domain assodated with kfB. There exist a semi-

distll'bution q and a random hetion q which is poIynomiaI on paverage such that q(z, s) -tj(x, s) 2 k(z, s)

and C(z) 1 tj({(z, s) 1 z E Q ( M , B, x, s)}) for aII z and z. We can assmne without loss of generality that

mmeg(z) > 0 for all strings x, and that on the empty input A, M does not make any queries.

For brevity, we write cT(z,rz) = Ij({(z, s) [ z E C A z E Q(M, B,z, s))) and &) = r> 1(z, n). Notice

that the sum c= 6(z) does not exceed 1 because

Moreover, we write Q(MT B) = U(,,,Er Q(M, B, z, s). To obtain the desired results, we let B' = {zOln [

z E B, n E N) and dehe d as f o l l m

G(z) i£ u = zOln and 6(z) > 0 for some n E N iY(w) = and some z E Q(M, B),

otherwise.

Smce i'(zOln) = P(z), we hiwe

Thus, Jbecomesadistribution

Fm we shd show that (H,4 (B,v). Fm an element a in (because of B # C). Deihea

function f as foII0ws: r

{ z ifw=rOln,

fW = a otherwise,

It is clear that f is P-computabIe, and f reduces B' to B- (Note that this f is not phonest.) For the string

a, i$ ( a ) = Y({w [ f (w) = a)) = 0 by the dehition of f, and thus we have ti(*) 1 i'', (lo)- For other

stringg 2 7

m

C(r) = 1 6'(dln) = iY(g I %dn[m = dlu])) = ~ ( { T u 1 f (w) = z}) = gP (r).

CHAPTER 5. AVeRAGE POLYNOMIAC TIME REDUCIBlLSIlES 178

We next show (ii) and (ii) of the lemma We dehe a new randomized oracle Turing madhe M' which

works as foIIows on input z, M' simnlates M on input x, a d whenever M queries a string z, M' queries

zOlltl to the oracle. This simulation is carried out on each computation path, Note that the random-input

domain associated with (Mf,B3 is the same as that with (M, B). By defbition, zOllrl E Q(M',B1,x,s)

if and only if z E Q(M, B,z,s). CIearly M' wi th orade B' compntes A with bomded-error probabilityty

Moreover, the length of a queried string is longer than that of the input. The rest of the proof will be

devoted to showing the domination condition for (Mr, B').

Now we introduce a random hct ion g. Debe

Note that 0 < g(z,s) < 1 for all z. In particular, g(X, s) = 1 because M does not have any query strings.

Using this frmction g, we also define 7/ as follows:

fir (2, 9) = fxz, 4 9(z, 4.

We shalr show that Y(zOIn) 2 7jf(((z7 s) I zOln E Q(Mr, B', 2, s))).

Chins 23 V(zOln) 1 f({(z, s) 1 zOln E Q(Mf, B',z, s))).

Proof of Cloinr For every z in Q(M, B,z, a), it follows that ij'(+, s) 5 j(z, s) - w. Hence, Tor every r

and n, we have

We next show that p' 3- q'. By the de6nition of$, it follows that

To show the desired resnlt, we should show that the random frmction l/g is polynomial on paverage since,

if so, Lemma 3 3 3 ensures that the fimction q/g is also polynomial on paverage Let us assmne that q is

p' on p-amqp , where # is an inQeaSiTLg polynomial For simpli-, write Q,, for Q(M, B, x, s) and let

Az)=e-#(k) - Weshattshowthat l/gisponp-average Assmner>I. Let&= {(z,s) € r [ z € C f h

T3ne$(z; s) 5 p(lz-1-3r))- Note that, for every (z,s) € E, if E(z) > 0, then Is1 5 Tirmefi(qs) 5 p(lxl - a ) -

Moreover, m this ~ s e , we have I(z,s)l I 2121 + Is1 + I < 2lzj(p(Jzj - 3r) + 2)-

CHAPTER 5. AVERAGE P O L Y N O U TIME REDUCIBEEZES

It is dear that the &st two terms are bounded above by 1/31., Let T, be the third term, and we &ail show

this term is a h bounded by 1/31.. It holds that

Notice that g(z, s) < 1, and thus, 6(z, I t [ ) > 0 for some r E Q.,. Then, g(z,s) = f, wme z-

5.5.2 More Structural Properties

We shall show another important lemma beIow- B&re stating the lemma, we prepare some notation-

Let M be a randomized o d e 'Ihring mlrhinP. and Iet N be a randomized lbring machine We dehe

another =domized 'Ihring machine MN which, on input z, does the hllowing: it simulates M on input

2; whenever M qneries r, shnhtes N on input z; and halts if Ad does. Let f' be the the random-mput

domain of this composite machine MN. Let z be &d For each random seed r E F(x), r, denotes the

assx&ed mdom seed generated by M w i t h some oracle on mput x, and r, denotes the associated random

seed generated by N on mput z.

Using this notation, the lemma is stated as MOWS.

C&QPTER 5. AVERAGE POLWOMC4.L TIME REDUClBlMIZ73S IS0

Proof. For brevity, we write TV in place of r M ~ . We choose an increasing polynomial p~ such that g

is on waverage. Smce p Sm I], take an increasing polynomial p and a random function q such that

q(x, S ) . G(z, s) 2 h r ( 2 , ~ ) for all pain (z, s) E I", and q is p on paverage.

Set

= PA (32) ' PA@A (32) ' 6z2p(3z)) f '3 t

where '3 = wE~(,) h(X, r). We shall show that h is fi on paverage. Fix d 2 1. Let Ed and E~ be the sets

de6ned by

Let Qr, denote Q(M, 42,s). We estimate the term k({(z, r) ( h(z, r ) > @&/ . r ) } ) as fonows:

Clearly the &st term is

SimilarIy, the second term is a h bounded above by 1/3d Let os denote the last term by Td and cafcnbte

Thus, T, is bounded above by the term

Skew avbpptt-reducii turns out to be closed under pm-reduability.

Proof. For simplicity, let us assume titat (Al, pl) SP, (A2, k ) and that (.4=, M) is skew avbpptt-redoable

to (A3,p3). Let f be an appropriate reduction which reduces (AI, p l ) to (AZ,R) with the domination

condition for f. Fbrther? let M be a boundedsrror pmbabiIistic oracle Turing reduction which reduces

(A2tP2) (-43t~33)- We shan consider the randomized agorithm N de!6ned as follows: on in@ z , h t compute f (x ) and

then simnIate M on hput f (2). For each 6 . d 2, since A1 (2) = A*df(z)),

This shows that N is a bodederzof probabilistic orade 'Ibring marhme with o d e &. The estimation of the nmning time of N with d e A3 on mpnt z with random input s, Time2 (z; s),

ismby Tme$ (G s) 5 c (Tie (z) + T i e s (f(z)) + 1)

for some constant c > 0. Notice that Xz-Timef (2) is polynomial on pl-averap, and Azs.TrmeS(f (z)), as

a random fimction, is aIso polpomiaI on pr-average by Lemma 52.4. In comeqaence, ~\zs-Tii$ (Gs) is

p o l ~ d on waverage. To see the domination condition for N, write t& for the defanlt d L E t r r i a t h indud h m and F;YA,.

Then,

For w, it holds that

This sum is taken over all pairs (x,s) € rN~S such that

This condition is equivalent to the condition z E Q(N, &, z , s). NOW let us define p(z, s ) = p ~ ( j ( z ) , s ) -pi (2)

and j(x, s) = f i ~ (z) 2-1.1 fp(z , s ) if s € rNAs (z) , and 0 otherwise. Notice that k-1 f (z)[ is polynomial on

pl-average, and consequently, k r s . p ~ ( f ( z ) , s ) is polynomial on pl-average Thus, p becomes polynomiaI on

PI-average. Then it follows that

This compIetes the p m L D

We need a relativized version of the AmplScation Lemma (Lemma 3.531).

Proof. Let us assume that (A, p) is skew avbpptt-redua'ble to (B, v) via a tandomized Thing madim M and a semi-distniution q. Let r be the random-input domain associated with M with o d e B. We suppose

that PrM[@(z) = A(=)] 2 $ + t Smce p 9 q, there is a random function q which is poIynomia on

p-average such that q(z,s) - ij(z, s) = h(z, s ) for all pairs (x,s) E r. Let be the minimal string that is

not in B. We can assmne that M does not query any strings of length mallet than or equal to Ial- Forthedesiredpmblem (B',r/), weset B={rOlnI Z E B A ~ E N ) andset

Gw(ln) - P(z) if w = dln for some n E N, g(w) = Oiherwife

Obviously d becomes a distribution became G'(rOln) = P(z) for dl strings z.

CHAPTER 5. AVERAGE POLWOMTAL TIME R E P U C I B D

We then show (i). Let f be

z i f ~ = z 0 1 ~ f o r s o m e n E N , f (4 =

.?o otherwise

The function f becomes a reduction function which reduces B' to B. To show that (B', d) sp, (B, v), we

should check if the domination condition for f holds. This is shown as follows:

Next we show (i). Let us recall the randomized aIgorithm in the proof of the Amplification Iiemma that boosts the success probability up to 1 - 2-lfl"m on inpd pair (z, lm). Here we slightly modify its dgorithm

to allow the algorithm to make queries. Take an integer c satisfying c > 1/c and let p(z, y) = @(Is1 flyl) for

all pairs (z, y). Then, we let N be the randomized o d e Turing machine defined by the following algorithm:

begin rand& algorithm for N

inpat (z, Y 1 if y 4 (1)' then reject f o r i = l top(x,y) do

sixdate M on input z without gaeries until M produces a query list

end for

Iist all possible query strings

fbr i = 1 to p(z,y) do

while shmhtion do

if M makes a cpery r then query dli

end while

ePd for

if the major& of the oatcomes is 1 then accept else reject

end.

By a simihr argument to that in the proof of the Amplification Lemma, we can prove (ii).

Notice that, for any string z and y and any random seed s, the reIation zOln E Q(N, B', z, y, s) implies

that y = lm and I < n I p(z, lm) for some uuxuber m E N Moreover, provided that I n 5 p(z, lm), we

have the dose reIationship

where each % is in r(z) and is assodated wi th s.

We shall show (iii)- Let fV be the randam-input domain asoc&ed with N with orade B'. Then, we

define rJ as f o l l ~ ~ ~ . Fix (z,Y,s) E F, I& US define b ( ~ , 9, S) to be 8(p(z, y) + 1)* - ziu) &,st) a d let

CaAPTER 5. AVERAGE POLYNOOML4L Tll\dE REDUCJBILITIES 184

Now we must check che domination condition for N; that is, iJ(((z, y,s) I dln E Q(N, Bt,z,y,s))) I 9(z01n) for any z and n. Using the previous calculation, for each string of the form zOlm,

The last term is equivaIent to the foiIowing term:

Therefore,

This compIetes the prooE

5.5.3 Bomded Error Probabilistic Thth Table Reduciiility

An ordinary botmded-ermr probabihtk oracle %ring madhe can rtiminish the error probabm signif-

icantly by repeating the same camp- and taking a majoriv vote to determine the outcomes of the

m & h Om~tedncti~~tdoesnotseemto~thisprapettpbecanseoftightdomination~diti~~~ifm

thedudion- Togaaranteeit,weneedamauy+metra * " asdasthesIommiuction.

Now we introdace onr barmdeckm probabitistic trrrth-taMe red-

De&rition 5.5.6 (Bounded-Error Probabilistic ThtbTabte Redactions) Let (A, p) and (B, v)

be distributional decision problems.

r. (kp) is bwnded-m pwbabdktk polynomial-time W-table redudble (bpptt-reduciile, for short)

to (B,u), denoted by (44 &P (B, v), if there erdsts a distnibutional decision problem (B', J) such

that (A, p) is skw bpp~reduaIbk to (B', v'), and (B', v') is pm-reducible to (B, v).

2. (A, p) is hde&enur pto- memge p o m - C i m c hth-h6& neducible (avbpgtt-reduable,

for short) to (B,v), denoted by (A,p) (B,v), if there adst a distnbutiod decision problem

(Bf, 4 snch that [A, p] is skew avbpptt-redua'bIe to (B', 4, and (B', v') is pm-reducible to (B, P).

In the following, t ~ e &OW basic properties of the dudbilities.

Proof. (1) By definition, detc, ' 'stic %5ng machines are a specid case of probabiIistic M g machines.

(2) The claim of dexiv i ty is obvions by choosing the identity redaction, ie., f {z) = z. (3) Clear from the d~~ (4) Here we that cPP is tran&be. -*g that (AL, PI) IP$ObPP (A2,pz) and (A27 p?) 5zbpp

(&t&), we s h f l sfmw (AI,PI) srPP ( A r ~ 3 ) .

Let (4, &) be a distributional problem such that (A1 , p l ) is skew avbpptt-ceduaIbIe to (4, A) , and

(4, &) is pm-reduciik to (Az7 &)- Suppose that (A2, M) is skew avbpptt-redncibIe to some distri%utionaI

probIem, say (A;, A), which is pm-redua'ble to &,~(3). By Lemma 5.5.4, (44) is skew avbpptt-

ducible to (A;,&). Appty Lemma 5.5.5 to this skew avbpptbrednction to obtain another randomized

d e Tking machine N, another semidistriion 7, and another distniutional problem (g,&) which

s & s k conditions (i)-(E) of the I ~ m m n as weJl as (A:,&) 9, (A3,ps)-

For readability and simplicits, m what fdhws, we on assume that (AL, pL) is sloew avbpptt-reduciile *

to (Az,m) via a Se2HEdtSttliutioo q and a bormded-error probabilistic o d e Tudng machine MI which,

with oracle A?, is polgnomid-time bounded on pz-average Moreover, ~ r , [ ~ p ( z ~ l ~ ; s ) = A2(2) I s E rrC(z7 lml] > 1 - TC'-, pz x U- Ps Q. B L W~({(~ .Y,S) I 2 L Q4(&,47~0. ,4}) for some ~i~~ anctsmnrandon&aade~machine M2 which,withorde&, mmin

polgn- time on waverage By the dominatian mditions, there are random frmctiong m and p2 which are plpnomial on m-average

and Pory~mial on ~2 x v--a- -, that fi (x, s) - i j l (x, 4 2 t r (4 - 2-181, P ~ G -9 -

CaAPTER 5, AVERAGE POLYNOMIAL TIME REDUCZBIIJTlES 186

begin randomized algorithm for N input z (say, n = 1x1) sinmbteMloninptttzrmtilthehstqueryismade complete a list of query strings

let m,* be the number of query strings

resume the simuIation

while simnlation do

if M2 queries z then sirdate M2 on input (2, I"'-=+~)

if MI reaches a halting configuration then output MI (z) and halt end while

end.

Let r be the random seed generated by N with oracle & on input z. Clearly random seed r is associated

with all random seeds generated by M2 with orade A2 on input strings which MI with o d e A2 queries.

For such random seeds, we write r, to de@pate the random seed that is used for the computation of Ml

with o d e A2 on input x, and d r= for the random seed for the computation of M2 with o d e -43 on

input z provided that z is in Q(M1, A2,2, rr ).

Given a pair (2, r) E , the nmning time of N with & as an o r d e on input z with random input r

is bounded bv

CHAPTER 5. AVERAGE POLYNOMIAL TlME REDUCZBIllTIES

It remains to establish the domination condition for N. For each z,

Write &(w, y, s') for p2(w, y, d ) - 8(IyI + I)*. Using the faa that 8[lyI + 1)* - ttrtlY(y) -> 1,

Also we have

Combining both inequalities, we can calculate the lower bound of b3(2) as folloftrs:

fil(4 2 - ~ l . 1 E m - -2-1'4

(t.4 &b? l-=+3,rw1 we..,

F d y we can show the dosure property of the dass Aver(BPP, *) under avbppreductions- This is a direct consequence of Proposition 5.544).

ProoE s'mn'lat to &at of Theorem 5-

5.5.4 Application of Probabilistic Reducibility

The motivation in having introduced the bounded-error probabilistic reduabiIity is to show that distrriu- tiond problems m Dist(NP, P-cornp) with nattd flat distributions become comp1ete. As a simple example?

we shall demonstrate that a h t version of the randomized bounded halting problem is truly complete for

Dist(NP, h m p ) with respect to bpptt-redua'bility.

Let us introduce a flat version of the rmuImnizai lkndai halting p b h (RBHP la) (BHPfta, C ( B H ~ , ~ . ~ ),

in the foflowing fashion:

BHP lclt = {(si, z, t ) / M. accepts z in It[ time }; and

begin randomized algorithm for N input (sit Z, lm)

generate a random seed w of Iength m

query (si, z, ta) to oracle

if (sit Z, w) is in o d e then accept else! reject

end.

Notice that there is no error 3 %HPfId is chosen as an oracle; that is, Pr,[lVBEIPf ((si, z? Im) ; s) = A(=)] = 1 .h N make nonadaptive queries and nms in poiynomial time for any o d e ,

Let c be an integer such that c 2 r/fiwd(si). Let us d&e the semidistriiution 7 as i)(r,s) = bHP (2) - 2-181 for dl pairs (2, s) E r- ; otherwise, 0. We tbezt have

Another significant application d the bppttreduci'bility given by -David et aL [9] is that

butionat NP search p r o b l d with Pmmputable distributions are actually reducibIe to distributional

decision pb1ems m DLPt(NP, Pamp). Nonetheless, sina we have not d e e d udIstnitrtionaI NP search

problems" and the main subject of this thesis is decision problems, we state the resnlt bdow without p d

The intererted reader ximy refer to [9] fix the deWi011 d search probIems and the proof d the theorem.

CHAPTER 5. AVERAGE POLYNOMIAL TIM23 REDUCIBIllTlES 189

p. TRen .thm is a &diu&d &ision probkm (D, v) m Aver(NP, P a m p ) such that (I?, p) is bpptt-

reducible to (B, v ) .

In womkase complexity theory? it is easy to see that if A 5% C and B 55 C, then A @ B C. The same is true m awmgeae eompl-. In the next lemma, we prove this claim.

RsealI from Section 3.5 the W t i o n of p @ v for two distn'butions C( and v.

Proof. We demonstrate only the case gPP. The other eases are somewhat similar. Let us assume

tbat (A, pA) 5Fbpp (C7Q via MA and ( B , ~ B ) <Ppp (C, () via MB. Dehe a new dgorithm N: on input

z, if z = Ou for some u, then simulate MA on input u; if z = lu for some u, then simulate MB on u;

otherwise, reject the input. For the sake of convenience, write v for p~ p ~ . Take a computational path y

of the computation tree given by N on inpat 2. Smce N basically follows either MAIS computation or MB%,

we can determine the unique computation path, which is made by either MA or MB, corresponding to the

computation path spedfied by random input s. We denote this computation path by 9. We show that ( A $ B, Y ) sbp (C, 0 via N. By om assmnptioq there exist semi-distriiutions q~ and

VB that PA 5- '?AT Ccg 9Bt i (2) 5 6 ({(% s) 1 2 E QWA, C7 u7 S) )) and &z) 1 6k({(% s) I 2 E

Q(MB, C, rr, s))) for all z, where IJA and IJg are the induced distniutions Eiom C(A and rMAc , and from p~

and I'G ? respectiveIy. Let q be d&ed as

$ - j ~ ( u ) if z = Ou for some u, i€z=luforsomeu,

otherwise.

It is obvious that q is a distribution. We also let q = q~ @ q ~ . Let

pA(u) if z = Ou,

otherwise

This is seen as foJIows. Assmne that z is of the form Ou. The

W y , we show that t(z) 2 ijr({(z, y) I z E Q(N, C p , y)) ) for all z. For each string z, we have

CBAPTER 5. AVERAGE POLWOMIAL TllldE REDUCIBlLITlES

This compIetes the p t

5.6 Structure of Reducibility

In this section, we shall take a dose look at the structure of the randomized complexity dass Dist(NP, P-comp)

in the Iight of rednciiility. F i we shall review the results of Ben-David, Chor, Goldreich, and Luby [9J.

They proved that if there is a hard distributional problem m Dist(NP,P-amp), there is also a problem

which is in Dist(NP, P-comp) bnt not pm-complete for Dist(NP, P-comp).

P m p o d b n 5.6.1 [9] Let p E strict-REC-comp ond D E REC. Assume that there QiPts a t h e -

corrPtrucfiMe function p such hat p(n) € no(') ond Ip(P(')) - p( la ) ) [ 5 2" + 2-1 for aftnost dl i, j € N If (D,p) 4 Aver(P, *), Men here ezirts a smri-diddution v such that (i) (D, Y ) 4 Aver(P, *), (n) (D, p)

(D, v), (5) (D,P) (D,p), and (itt) Y is computed by an oracle log-space detffministic Timhg machine

n l a t i o e t o t h e ~ o r n J e p .

Proof. The proof proceeds by a d e d slow diagon&don technique. We first enumerate aII quadruples

(Mi, Ni, ci, 4) of a deterministic Turing machine Mi, a determiniStic orade ntring machine N., and natural

numbers ci a d 4. Write also Di for L(&) for brevity- Note that by OW assumption, there exkts a

dekmhi&c Turing machine which exactly computes ji.

For a finite sequence a and an integer k 2 1, let a(k) denote its kth element of a. For a b i t e sequence

u of 0's and I's, Iet

It is easy to see that, for 4 x, fi(x) = 0 if u(Iz1) = 0, and P(z) = i(z) otherwise. Note that the condition

on p that Jp(P('3) - p(lm)l _< 2" + 2-j will be used Iater to goarantee the existence of a "normah&

distribution of Y.

Let us consider the fonowing subprogram Cond(i, a, t):

subroutine Cond(i, a, t)

if t steps are consumed dm* the following computation then return %on

if i = then let i, r= Li/2] and go to (*)

let 6 = [(i - 1)/2J (since i = "odd")

check the fdowing two conditions

(i) D, n C ~ I ~ I # D n Slul; and

(ii) Cso<ltl<lol - TimeMio (x)ll& - 121-I - fi(z) > c, if either (i) or (ii) is true then retmn "yep" else return %on

(*I check the following three conditions

(i) L(N,, Db) n SIUl # D n Glul;

CHAPTER 5 AVERAGE POLYNOMIAL TIME REDUCIBILITLES 191

(ii) ~c,zl<l,,l - Tim:: (z)'/h - - ji(z) > ~ q ; and (i) there exist two z, y of length < lul such that ji(z) > 3(y) = 0, and y f Q(N& Di,, ,x)

if either (i), (ii) , or (iii) is true then return "yes" else return %on

Notice that if Cond(i,o, t) answers W,'' then so d m Cond(i, T, t) for all extensio~ T of u (i-e., r _1 u).

The following is the main body of the algorithm which computes a:

begin deterministic algorithm for u

input k

let t = pgkJ

f o r i = l t o t d o

exit the for-loop when t individual steps are executed in this for-loop

compute 41') ifp(li) > 3 then go to (*)

end-for

output u(k) 7 I and haIt

(*I f o r j = l t o t d o sit the for-loop when t mdividd steps are executed in this for-loop

recompute aCj) (by a recnrsive On)

end-for

let J be the Largest index j for which u(j ) has been recomputed m the above for-loop

if no such J exists then output a@) = 1 and halt

set FJ s o(1)4(2) . .*u(J)

f o r i = l t o J d o

exit the for-loop when t individual steps are executed m this for-loop

call Cond(i, F J , t)

if Cond(i, FJ, t) = 'ho" then exit the for-loop

end-for

let 1 be the S I d & index i f0r which cond(i, FJ, f) anSWerS %on

output u(k) P 1 mod 2 and halt end.

The exit-statement enforce a time bound t (= LlogkJ) on each each for-loop; thns, this aIgorithm

req@res at most O(Iog k) steps. Hence, u is computed by rising 00ogk)-space. By definition, u is correctly

mputab1e using 1%-spacewith thehelp ofp as an orade-

We shalI show that this algodhm c o d y computes o(k) on input k. We assmne that there is no

extension T of u such that Cond(i,r,t) answers -'' Assmne that i is odd. Then r is of the form dm

for some m E N because a(j), j > k, is atways set to be 1. We then have D, n e1'1 = D = 6 1 7 1 . This

implies L(Mh) = D. .b we have LM<ltl - Theas, (x)~'% - - 9(2) 5 G. We remadc that the

d u e P(x) equals ji(z) if 1x1 > 101. Hence, for some constant c > 0, we have

This implies that (D, p) M o w to Aver(P, *), a contradiction.

Now let us consider the other case that i is m. Notice that any extension r of u is of the form dm for

some rn E N. Then L ( N , , D,)~CI~I = DNZ~I~I for all extensions r ofu. Thus, L(Nk, D,) = D. Moreover,

there are no pairs (x, y) such that b(x) > 0, P( y) = 0, and y E Q (N, , D, , z) . RecaU the definition of u. Since

7 = dm, rn E N, it foIlows that, for any y of length great^ than lal, 5(y) = 0- This means that there is no

z (121 > lo[) on which N . queries strings of length greater than lol. In other words, 11 U, Q(Nh7 D,, x) 11 is finite. Let m dehe another lhing marhine N' which sindates Nh on the same input with dl orade answers

born Dh being encoded into its program. Clearly L(N,, D,) = L(N), and Tme~(x) c ( ~ i m e 2 (z) + 1)

for some absolute comtant c > 0. Notice that &,<ltl<lTl - 121-I -Tune:: (2) '1% - j i ( t ) c, for all extensions

T. Hence, we have

This again condudes that (D,p) is in A m p , +), a contradiction. Thdore, the algorithm for o works

PW=IY - By Corohy 43.3, we can normdize v to a W distribution u' such that B' (2) = P(z) for all nonempty

strings 2. 0

kr a simiIar fashion, we can demonstrate the existence of a distniutional problem that is not pT-harder

than a given problem, bnt can be pm-reduced to the giTlen problem,

h o f l Similar to Theorem 56.1. a

Root Assmne that Dist(NP,P-comp) P Aver(P, *). Then, no pmamplete probIem for

Dist(NP,P-comp) belongs to A*, *). Let (D,p) be any anY'buticmaI problem which is pm-complete

h r Dist(NP, P-comp). SinQ (D, p) # A-, *), Propositim 5.62 gmar&es the exkfience of a set E such

CHAPTER 5. AVEXAGE POLYNOMIAL TIME REDUClBILITIES 193

We say that two distriintionaI problems (D,p) and (E,v) are incompmable with respect to 5; if

(Dl P ) g (6 and (E, 4 2g (D, PI- The foIlowing theorem leads to the existence of incomparable pairs in Dist(NP, P-comp) with respect to

pT-reductions.

Proof. We use the dow diagodhtion technique again. Let {Mi)iEN be a standard enumeration of

aU deterministic polynomia-time o d e Turing machines. We recatt that si denotes the ith strtog in the

standard order on C'. Let m,r be idkite segaences of 0's and 1's. Let i (z ) = 2-21rl-L f ~ r all z . This

distribution is P-computable- We next define a semi-distribution p as

for any nonernpty string z- S i l y , v is defined by repking u with T. Note that u(lz1) = ~(121) if and

only if jqz) = P(x).

The infinite sequences a and T are computed by the fbllowing recursive procednre:

begin deterministic aIgorit6m for (u(n), r(n))

inpat n

i f n = 0 then output (u(n),r(n)) s= (1,l)

Iet t = [Iog nJ (t does not change dnring n E {P, T + 1,. . . , 2m+1 - 1))

for j=O to tdo

exit the for4oop when t steps have been executed

==P* a P* ( ~ l j ) , dj)) @Y a ~~- cam

end-for

let J be the Iargest index j for which (u(j),r(j)) has been recomputed

i f no such J exists then output (u(n),r(n)) = (1,l)

for i=l toJdo

d Cond(i, J, t )

if Cond(i, J, t ) = "no" then &

d f m

let I be be tbe mdex i for which Cond(i, J,t) = "no"

(fmthesakeofcosmenience~, writerfn) =I)

if no such I exists then output ( ~ ( n ) , d n ) ) = (1 , l )

output (u(n), ~ ( n ) ) = ((I + 1) mod 2, I mod2)

end.

The subroutine Cond(i, J, t ) is as follows

subroutine Cond(i, J , t )

if i = "odd" then let io = L(i - 1)/2j and go to (*)

let io = Li/2J (because i = "evenw)

check the fdowing two conditions

(i) L(M,, D) n EslPI # D n 2Slol or

(in) there are x, y < la[ s.t. ji(x) > fi(y) = 0 and y E Q(M,, D,z)

if either (i) or @) is true then mtum "ges" else return %ow

(*I check the folIowing two conditions

(i) L(Mb, D) n ~ l l ~ [ # D n ES[V~ or

(in) there are z , y < [a1 s.t. 32) > ji(y) = 0 and y E Q(Mh, D,z)

if either (i) or @) is true then return "yes" else return "now

return.

Pnwf of Claim. Assume ran&) # N Take the minimd integer I such that r(n) 5 I for all n > 0. Since I does not change, the algorithm always takes the same index I. F i consider the case that I is even, Let io = 112 and let n be sufEciently large. Notice that, for every

z of length at least n, ji(z) > 0 and G(z) = 0. Moreover, for every z, we have ME(=) = D(z) , and ME(%) does not query any string y, where ji(z) > ~) = 0. Hen* ME computes D on all inputs, and it queries

only strings of length smaller than n. This i m p h that D is in P, which contradicts our assmnption. The

same argument also holds for the case whem I is odd I

Therefore, p and v are well-defined. Finany, we must normalize these two semi-distn'butions to obtain the desired distributions. It is not hard

to see that p and v are log-space computable since, m each stage n, we qtdt the simuIations after Iogn steps

are made- Ekom Corollary 4.33, it it to toow that, for p(n) = n + 1, lp(ldg) - p(lP(-f))j 5 Ti + 2-1

holds for almost all i , j E N. Assume that i 2 j . Then,

CHAPTER 5. AVERAGE POLYNOMIAL TIME REDUCIBxLxTIx.3s 195

because either ji(Ck) = *(EL) or j i (Xk) = 0. By Corollary 43.3, p can be normalized by p' such that ji' (2) =

ji(z) for an nonempty strings z. Thus, p' is Psomputable- In a Jimilar way, we can show the existence of

the P - c o m p d e distniution d which normakes v. Dishiutions p' and J satis@ the conditions of the

theorern, Thus, we complete the proof. 0

Proof. Asmme that Dist(NP, P-comp) g A m p , *). Let D be a set in NP - P. This set D ackfs

because P = NP implies Dist(NP, P-amp) E Aver(P, *). By Proposition 5.6.4, there exist two distribu-

tions p, v E P-comp such that (D, p) and (D, v ) are incomparable. Notice that (D, p) and (D, v) are in

Dist(NP, P-eomp). 0

5.7 Recent Topics

In this section, we shan discuss some issuis which have emerged recently.

Randomized Many-One Reductions. The search for a better definition of "reductions" is an exciting

field m averageease c o m p l e theory, because the Werent choices of reduciiility can lead to different

worlds. -

For example, we have seen that there is no Bat distrrriution for which a distrritational problem is pm-

complete for Dist(NP,P-amp) if EXP # NEXP. However, Gnrevich introduced a random version of

many-one reductions under which m e Bat distriiutions make their distributional problems "compiete" A

shnilar phenomenon occurs m the case of random reductions. Impagliazzo and Levin [44] demonstrated that

distn'butionaI NP-search proMems can be mdomly reduced to distzibutional decision problems.

Here is the dehition of random redttctions. Recall the debition of dilations from Section 3.6-

DeaPition 5.7.1 [l2l For two dishiiutiona1 problems (D, p) and (E, v), (D, p) is polynomial-time ran-

dody (or randomizing) reduai1e to (E, v) if there exists a fimction f which is mmpdable by a randomized

aIgorithm m poIynomial time on paverage, and a dilation r such that (i) for any (x, s) E r, z E A if and

ody if f (z, s) E B, and (5) f satisfies the domination condition. (NB. f takes inputs of the form (z, s).)

Using this type d reduction, the fbHowing interesting daim can be proved.

CHAPTER 5- AVERAGE POLYNOA5A.L TIME REDUCIBjLR7ES 196

AIternative Dehition of Polynomial on the Average. Let us take a quick look at Levin's definition of wlynd on the a m n R e d l that a fimction g is polynomial on pawrage if and only if z.,C* fi(z) < 00. One of the obsavations is that it allows us to deaI wi th distniutions that put too much weight

on the &st f m strings. J. Cai and A. SeLnaa [I91 considered such distri'butions inappropriate for a coherent,

generat theory of averagecase complexity. As an alternative, they propose a diffefent notion of " p o l y n o d

on the avgage" which counts on the convergence rate of the expectation.

D W o n 5.7.3 [I91 A function g is "t on the paveragen (in the sense of Cai and Selma) if, for aU

n L 1,

As long as p d e s An-b({z [ It1 1 n)) E R(n-k) for some constant k > 0, Levin's definition of

"polynomial on p-avefagen coincides with that of Cai and Selman [19]. Moreover9 this notion of Cai and

SeIman is equivalent to the notion of "polynomial on average with respect to {ph),&' - in Definition 33.7.

C. Rackoff aIso suggests (reported in [19]) the following definition-

D e n 5.7.4 [In] A function g is "t on the paveragen (m the sense of RackofT) iE, for a11 n 2 1,

Rackofffs definition also provides the same notion of "polynomial on paverage" given by Levin if p

satisfies h K { r I I%[= n)) € Zr(n'k) for some constant k > 0.

However, Cai and SeIman7s d w o n seems inappropriate when we consider the polynomial-time many-

one reductions introduced in Section 5.2. Let us denote by CSAverP, T ) the collection of all distriiutiona1

decision problems (A+) such that p E F, and A is computed by a detemxhbtic %ring machine M which

safSes th condition that, for aU n > 0, &, w j l ( x ) 5 fi({z I It1 2 n}) for some constant

k>O.

Belanger and W q [?'I showed that CS-Amp, *) is not closed under S ~ - r e d u c t i o ~ or ~F-reductions

with o n m e reduction functionsflS Hence, CS-Aver(P, *) is properly included in Aver(P, *).

Chapter 6

Average Case Hierarchies

6.1 Introduction

One of the most novel contributions to worst-case compl- theory is the introduction of the p o l y n d -

t h e hierrrrJry (or poipmid k z d i g ) by M e p and $toclaneyer in 1972. This worst-case hierarchy is

built &om P and NP m a m y similar to how KIeene constructed his arithmetical hierarchy above recursively

enmnerable sets. Notably, the classes of the hierarchy Iies between P and PSPACE. The c~mtmction is

such that the kth I d of the po1ynomial-time hierarchy is d&ed to be the conection of all sets, each of

which can be recognized by a polynomial-tirne deterministc/nondeterministic oracle Thing marhine relative

to some sets in the (k - l)th Id of the hierarchy.

The polynomial-time hierarchy has been studied exhsidy for over two decades with many lntrigtring

resttlts. For example, if the BooIeaa hierarchy over NP coIlapses, then so does the po1ynomial-time hierarchy

[4q, and if the polgnorniai-time hierarchy coIIapses, then the Iow and high hierarchies in NP coRapse [89].

Many NP-hard probIems are chssified into wariors levels in the poIynomia1-time hierarchy-

This chapter wilI b d d a- analogues of the po1ynomi;il-time hierarchy among distnhtional

dedsion prob1ems1 To introduce an average polynomia-time hierarchy, we W begin with the relativization

of the fundamental a9erage-case c o m p I w classes Aver(P, 7) and Aver(NP, F) , relative to a distriitrtional

problem (E,u)* A relarivization ofthe dass Am(P,F) to an orade probtem (E, u) is nattaany induced h n the average

polynomial-time dete ' ' tic Thing reducibiIity defmed in Section 52. A s b n k approach toward a rela-

tivized Aoer(NP,F) is taken by Schder and Yamakami [91 based on the model of docked nondeterministic

lhring xuachines. Here we shall d&e a &athi& Aver(NP,F) in a slightly different way. In Section 5 z

the relativiz;ition of Aver(P,7), Aver(BPP,F), Aver(NP,F), and Aver(PSPACE,F) will be formdated

m terms of various restfictions on average poIpnomia-time oracle 'king machines together with weaker

domination conditions for those machines.

In dativized worlds, we can see a deskable separation d Aver(NP, Pamp) and Aver(P, Pcomp), and

CKAPTER 6. AVERAGE CASEEIEWBm 198

a cohpse between Aver(P, P-comp) and Aver(PSPACE, Pcomp).

In Section 6.4, using lhring redudbiky, we shall introduce an average-case version of the poIynod-

time hierarchy, average polpnosnial-time hierarchy under a set 3 of distributions {Aver(AL, F), Aver(EE, F),

A=@i733 I k > 0). Another effective way of charaaerizing the kth level of the polynomia1-time hierarchy is to use poIynomia1-

time alternating Turing machines wi th k-dternation. This worst-case characterization is trery suggestive

and enables us to build another of average polpornid-time hierarchy using "averageD polynomial-

time alternating 'king machines with constant-alternatioo, We call this a- hierarchy the aaffoge

polylomial-time u h m a h g hiemdg. Section 6.5 wiH formally introduce an average-case alternation hd

erarchy based on average polynomiaI-time alkmating lbring machines with constant-alternation. As the

reader may perceive? however, these two types ofa- hierarchies are unlikely to coinade.

We have seen in the previous chapter that if Dist(NP, P-comp) is not included in A m p , P-comp), then

there are distriintional problems whieh are in Dist(NP,P-cornp) but which are not pm-complete. This

indicates the possibility of a large gap between Aver(P,P-comp) and Aver(NP, P-comp). One approach of

refining this gap indves constructing a hieratchica structure within Aver(NP, P-comp) and sorting out all

distributional problems in Aver(NP, P-comp) into various IeveIs of this hierarchy. This approach was taken

by Sch6ning m w o m e s e complexity theory in 1980, and the hierarchies within N P are d e d the high ond

low iriermdries. A n a d averagecase -on of the Iow hierarc&y, d e d the menage loto irienrrclry rrrithin

Aver(NP,F), wi l l be introduced.

Major Contri'butions. In this chapter, we introduce severaI new hierarchies.

In Proposition 62.8, it foIlows by the self-redua'bihy of RBHP that Dist(NP, P-comp) A m ( B P P , *)

implies Dist(NP, P-comp) G Aver(RP, *). The notion of relativizatian originany comes h m SchuIer and Y ' [951; however7 the relativizations

of the classes Aver(NP,3) and ,4ver(BPP, F) treated in this chapter are dSet-ent fiom those in [97j.

Proposition 63.8 shows that Aver(BPP,F) reIative to Aver(BPP, *) collapse to Aver(BPP, *) for any

set 7 of distributions.

Proposition 63.11 ~ ~ Q W S basic properties of relativized Aver(NP,7), such as re0- and transi-

tivity. In particular* if (A,p) is m A m , *)(By), then aIl problems in Aver(NP,F)(A') belong to

A v ~ ~ ( N P , F ) ( ~ Y ) .

Theorem 63.12 demo~~strates an o r d e separation between Aver(P, P-comp) based on a tally orade con-

struction given by B a h , Gill, and SaIovay [3], whereas Theorem 63.17 shows a collapse of

Aver(PSPACE, P-comp) to A*, P-eomp) m a reIativized worId. The proof uses a d a t h i d -on

of the randomized h d e d halting pmbIem.

Ropasitim 6- I o w s that Auer(PSPACE,F) reI;ltioe to Aver(PSPACE,+) collapses to

AmfPfWACE,T) forangset3ofdistribntions,

Basic indnsion telatianshipq such as Aver(AL, 3) G AWE;, F) G A m ( A ~ l , F), among cksses in

the average polynomiaE-time hierarchy arr shown m Propositim 6.44.

CHAPTER 6. AVERAGE CASE BIERARCfIlES 199

Propusition 6.4.6 shows that a collapse of two levels of the average polynomial-time hierarchy canses anp

higher level of the hierarchy to collapse-

An Important new idea is the use of of alternating machines to build another average-<.ase version of

the polpnomiaI-time hierazchy. Theorem 6.52 gives an oracle characterization of the average polynomial-

time alternating hierarehg; for example, Aver(AAh,,F) is equivalent to the dass Aver(P,F) relative to

Dist(CE, *). A similar characterization holds for Aver(ACR,, 7).

In Propcsition 6.5.4, it is shown that the dass Aver(BPP,F) is included m the second I d of the

average polynomial-time aIternation hierarchy.

Proposition 6.43 and Theorem 6.411 show that many basic average complexity dasses, such as A m p , *)

and Aver(NP, *), have the sparse interpolation property.

6.2 Distributional Polynomial-Time Hierarchy

In the previous chapterl we showed that sweral important distributional decision problems, such as the

bounded halting probIem RBHP, are pm-complete for Dist(NP,Pamp). This section win extend these

completeness resnIts to the class Dist(Ci,P-comp), the kth I d of the "distributional polynomial-time

hieratchy under P-comp."

6.2.1 D M o n of Hierarchy

Randomized complexity dasses are the simplest randomization of existhg worst-case complexity classes. .Is

with the polynomial-time hierarchy, we can comider its nattrral counterpart of &e polynomial-time hieratchy,

the d i d d d m d pdJlnomial-time hierorclry, under a given set of djstriiutions.

We begin with its format dehitiort.

Delinition 62.1 Let T be a set of distriiatio~~~. A d&ibt&d polyunnial-time KefPrthy tu&r 7

consists of the following classes Dist(Ai,F), Dist(CE,F), and Dist(II:,T) for all k >_ 0.

We next demonstrate the existence of pm-complete problems for each dass Dist(!EE,P-comp). We

generalize the randomized bounded halting problem RBHP, seen earlier? to the kth level randomized bounded

halting problem RB*. The kth leaJ rrrndomLzed boanded hdting problem RBHPI is the distriiutional

problem BE@,^^) defiaed m the foIIowlng fashion. Let c r ~ ~ p be the same distribution as in Seaion

53. Assnming that (Mi)ia is an e&tive enumeration of aIl nondeterministic oracle lhring machines, we

define

BHP(A) = ((i,z, In) I Mt accepts z m less than n steps ),

and then set B@ = BHP(0) and B*' = B H P ( B ~ for k 1 I- Next themem proves that, f i r each k, the problem RBHP* is pxmcompIete for Dist(CE,P-comp)- Note

that thecase k = l has been aIreadyshownasmTheorem53.2.

CaAPTER 6. AVERAGE CASElW3RA.R-

Roof. The general case k > 1 is very sidar to the base case k = 1 of Theorem 5.32. We note that

B@ is C~-cornpIte.

For every set D E XE and every distriiutioon p E P-comp, we shaU show that (D+) SP, ( B & , ~ ~ ) .

Notice that BIIP~-' is pwompIete for CE-,, and as a consequence, there exists a polynomial-time non-

deterministic orade 'I\uing machine M computing D with o d e BHP? Let g be the function g defined in Lemma 4.2.7(2). The fimction g satis6ies ji(x) < 2-19(')1+*. Now let i

be aa index such that L(Mi) = L(M). Let p be a polynomial time bound of Mi. For the reduction f from

D to B@, we set f (z) = {~ ,g ( t ) , l~(l'I)). Clearly f is oneone and witnesses the reduction D <k BEP~

because z E D holds exa&y when Mi with o d e BEE'&-' accepts z within p(jz1) steps- To check the

domination condition for f , we simpIy follow a argument similar to that used in the proof of Theorem 5.3.2.

nus obtain (D, P) SL ( B e ? mm)- D

Meyer and Paterson [7l] have introduced the notion of ( p o l y n d - t h e %ring) s d f - d t ~ c i b i l i t y into worst-

case complex& theory.

AIl known NP-complete problems are self-redua'bIe, and every self-redua%le set belongs to PSPACE. It

is natural to ask whether the notion of self-redum'b'i has a counterpart m distn'butiond decision problems.

We shall indeed give in this section the f o n d d-on d an averageease version of self-redua'b'tlility and

show the existence of self-reducii1e sets m each I d of the average p01ynomi;rl-time hierarchy.

We begin by dehing the important concept of a P-eornputabIe, OK partial order on the set F.

Dehition 6.2.3 (OK Partial Order) Let < be a partial order.

1 A partid order < is ~~~tinre cmpdabk (P-computable, for short) if there exists a polynomid-

timedeterminirtc~maehineMsneh that,forewppair (z,y),x < y ifandonly ifM(z,y) = 1.

2. A partial order c is OK if there akts a polynomial p such that

For exampIe, let us write x < y to mean Iz[ < 191 for any pair (2,~)- Then the relation < becomes a

P-eomputabIe, OK totid d e t on C.

The notion OfselErednciiilitp is now stated as f m

CHAPTER 6- AVERAGE CASE HIERARCHIES 201

Definition 62.4 (Self-Redna'bility) [97j A distriiutional dedsion problem (D, p) is (poEyunnial-Eime

ltting) selfreduci6k if there exist an OK partial order and a deterministic oracle Turing machine M such

that (D,p) 5: (D,p) via M, and for every mput z , all query strinp m the computation of M on input z

are smaller than z with respect to the partial order. The machine M is CaIled a self-reducing mcJrine for

( 0 7 P I -

Proof. (1) For any distributionaI probIern (D, p) in Dist(P, *) , there adas a deta ,tic htring machine

M which computes (D, p) without any queries. This machine reduces (D, p) to (D, p). By the definition of

self-reducib'ity, (D, p) out to be self-redualble (2) Assame that (D, p) is &-reducible. Let as take a self-rednurtg &e M for (D,p) . We shall

dehe a 'Ihring &e N which computes D using polynomial space. Below we describe an algorithm for

Nusingarecumbed

begin detE - ' 5c dgorithm for N

input z

s i m h M on mput z

while 5alation do

(*I ifdd r x d s a query z then simulate N on input z

if M reaches a hatting configmation then output M ( z ) and halt

end-while

end.

We must prove that N ases only p o 1 y n d space- Let us analyze the query process of N on input x.

Notice that the space used by the machine M is pbomded, and thus there are at mast exponentially-many

Werent co-. Remember that this number is independent of the choice of oracle. Let us assmne

that m the first r m d of the simnlation of M, M queries at most exponentiany-many striqs, which are in

Q(M, D, 2). The recursive protocol m the line (*) brings another round of the sirnnlation of M on each input

taken from the set Q(M, D , z ) . Fm an a r b i i striug dl) in Q(M, D p ) . h the second round, M makes

another set of gneries, Q(M, D, z(')). To go into the third round, we EK a query string d2) m Q(M, D, z('))

and then consider the set Q(M, D, d2)) of quexy stthgs. Recnrsively, we keep E&g a query string r(3 and

then stapt another round of the simnlaton of M on the input z (0 . This process proceeds until M does not

querpany- Now let us d d e r an arbitrary sequence z ( ~ ) , d 2 ) , . . . , z(*l, - - - taken by the above procednre Smce M

is ad-reducingmlrhinp. thissapncemakesastrictIy descedhgchain with respect to the given OK partial order. This impIies that the length of the dmiu is bounded above by a poEynomiaI m lzl, and the

Iength of ea15 string z('2 is a h bounded by a polynomial m 14. Let p be such a polynomial As a resnlt, N

CaAPTER 6. AVERAGE CASE HIEELARCHIES 202

caa simutate M using p(lzl) blocks on a workug tape, each of which is used to store one of the strings t('7

being queried in each ith round. Therefore, N needs only polynomial space 0

Moreover, the set of all self-reducible problems is closed under pisomorphism; i.e+ if (D,p) is p isomorphic to some distr~bntional problem which is self-reduciile, tben (D, p) is is-reduable

Lemtm 6.2.6 Let (D,p) and (E,v) be any d w h b d w d . . - problems. If (D,p) is pisomorphic to (E, v)

and (E, v) is s e l f - d d l e , then (D, p) is self-reducible-

Roof. Let f be a pisomorphism from (D,p) to (E, v), and let M be a M-reducing machine for (E,v). By the definition of pisornorphism, f is P-computable and pinvertj'ble; that is, f 'I is also P-computable.

smce (D,p) 5; (E, v) via f, there exists a polynomial p such that p(lzl) - 3( f (2)) >_ $(z) for a11 z. SrmilarIy, f'l reduces (E,v) to (D, p), and thus there is a positive, strictly increasing p o l p o d q such

that dlfb)l) - i(z) 1 P(f (dl for d z- We wish to construct another self-reducing machine N for (D, p). Let as define the machine N as follows:

begin deterministic algorithm for N with made

input z start the simulation of M on input f (z) while simulation do

if M queries z then query f -' (t) to orade

if M reaches an accepting configuration then accept

end-while

a~cept

end.

Recatt that D(z) = E(f(z)) and D(f-'(2)) = E(z) for ail strings 2 and 2. h r n these equations, N can reduce D to D in polynomial time. Ehahermore? t h a tAsts a strictIy increasing polynomial q' sucb that

if E is a query string made by N on input z, then If (z)l 5 q'(lzl), since f and f" ate both phonest and

pbomded Next we shall check the domination condition for N. By the domination condition for M7 we can find a

semi-distniution < and a polynomial p' such that p'(lx1) t(z) 2 O(z) and P(z) 1 &{z I z E Q(M, E,z)E)

for d x and z. We dehe the semkWriitztion q by G(x) = &f(z))/s[lxf), where s = q o 6. We first show

that p 9 7 - This is seenas foII0ws- Foreachz, w e b

b these inesnalities, we set t(r) = p(z)p'(f(z))s(z), and thus t(z) ij(z) 2 p(z).

CHrLPTER 6. AVERAGE C,4SE EERARCHIES 203

Th tam xrlEP(IVOT, kf(x)) b egnivala to the &{Y I f (4 E Q(M,E,Y)}, which i~ bounded by

D(f (2)). Therefore, it foIIm that

This compIetes the proof. 13

One of the dassid &-redua'ble NP-complete problems is the satig%biZity problem, SAT. However,

we do not know a simple distriintion p snch that (SAT,p) is Disc(NP,P-camp)-comp1ete. M o m , as

shown by Ranm and P a d [q, SAT is deterministically compntabk m polynomial-time on the average

under some natural distriion. Instead, we consider a skew version of the kth level randomized bounded

halting problem, &ewRBEP$ which is pisomorphic to RBHP'. We waat to show that skewRBHPk is

self-reduciibie. By Lemma 62.6, this implies the sdf-redutibii of RBHPk-

F i we W e the kh Id randomized skew bounded halting problem skewRBHPk as the distriiutiona1

decision probIem (BHPak,, e,,,). .4smming that all nondeterministic 'Ihring machiues take at most

two nondeterministic choices at every step, BHP,,, is the collection of string of the form (si,x, y, l t) ,

[yl 5 t, such that, on iuput z, M$ with oracle BHP~-' dm ' ' t i d y follows the computation path

specified by y duriug the ht lyl nmdeterminish'c choices and then halts in an accepting state within t

steps. The distribution map,, is de6m-A bpr

h o t As mentioned abme, our strategg is to show that (i) skewRBHPk is pisomorphic to RBHPk, and

(ii) skewRBEPk is self-reducible.

We first sketch the prod of statement (i). The p d that sIPewRBHPk RBHP' via a P-computabIe,

plnvertible reduction hc t ion is similar to the proof of Theorem 5.35. The other direction FtB@ 5; &WRB@ is shown by considering the reduction hctian f defined by

It is easy to see that the dation c is P-ie and that it is an OK partial order becanse of the upper

botmd [t'[ on the size of y.

Next let us consider the fo11owing oracIe lhrhg machine N.

begin det ' ' d c algorithm for N

inpd (i, z, 9, t )

if t is not of the form 1" then reject

if 181 > It1 then reject decode the code si and rerover the mrrhine Mi simulate Mi on input z deteminMcally foDowing the oompntation path

encoded by y until the machine either exhausts [ti steps or

makes the (ly( + 11th nondeterministic choice

if either Mi reaches a halting state or it does not halt within It1 steps then go to (*)

query (4 2, YO, t ) and (4 2, Y 1, t ) to o d e

if one of the strings belongs to oracle then accept else reject

(*I if M, accepts z within It1 steps then accept else *ect

end.

We wish to show that N is self-reducing machine- Clearly N reduces BEW~, to BHP~~ , by qnerging

only strings which are Smaner than input with respect to <. To complete the proof we must check the

domination condition for N. Notice that two query strings (si, z, y0, t) and {sit x, yl, t ) uniquely e~l~espond

to input (si, z, y, t). Let us consider the query string (st, z , yO, t). The probability ~ ~ B F K P , ~ (sir 2, YO, t ) is

bOu11ded

since llog(n t I) 5 Ilog(n) + 1 for all n E N. A s i m k ineqnality holds for khc.. (si, x,yl, t). Therefore,

N is a self-reducing machine. 0

As stated before, Wang and Betanger [lU] show that most known distriintionat problems compI& for

Dist(NP,P-comp) are pisoruorphic to each other. Fhm Theorem 62.7, we i m r n W y conclude that

most known pm-compIete problems for Dist(NP, P-comp) are seif-redudbIe. Nevertheless, we do not know whether aZ2 pm-compIete problems for Dist(NP, P-comp) are self'u~'b1e. If the Isomorphism Conjecture

(m Section 53) is tme, then all distdbutid problems in Distw, P-armp) obvionsIy become &-reducible- F d y m this section, we shan demonstmte an application OfseIf-redtxcibili~~ In mm-case complex&

theory, KO [52] and Zachos [I24 use d-reducible sets to show that NP C BPP impIies NP = RP. A simihrargamentcanbecatfiedontiutheaveragpcasesetting-

P n q w d b n 6.2.8 Dist(NP, P a m ) E Aver(BPP, *) if I#rd only if Dist(NP, P a m p ) S Aver(RP, *) .

Roof, Let us assume that Dist(NP, P a m p ) is induded in Aver(BPP, *). We have seen that skeftrRBHP

is sekeducible and compIete for D i s t w , P-comp). It is thus enough to show that &ewRBHP belongs to

Aver(RP, *) because the dass A m , *) is dosed under pm-reductions.

Let M be a seKred~ing machine for skewRBHP which nms in po1ynomiaI time, and a P-computabk,

partial OK order 5 for self-reducibty, Let p be a poiynomial such that TimeM (x) 5 p(lz1) for all z. Let us

consider the self-reducing tree of sk3ewRBHP. By our assumption, there is a botmded-amr probabiIistic 'hr-

ing &e Mo which recognizes BHP in po1ynomial time on mwaverage. Ben&ed by the AmpIification

Lemma, we can assume that PrMo[Mo(z) = BHP(z)] 2 1 - for aII z-

On input (S~,Z~Y, Im), each node of the seif-reducing tree starting fiom the root (si, z, lm) desc r i i

a nondeterministic choice made by the machine Mi on input 2. TO determine the correct outcome of the

machine, we probabilisticalIy trace -he self-teduing tree dong each path as a series of nondeterministic

choices; when we reach a leaf, we check whether the machine reaches an accept@ configuration. This Iast

step is done without any error, and thus we have a one-sided error randomized algorithm-

ThereforeJtBHP is in Aver(RP, *). 0

6.3 Relativization of Average Complexity Classes

In 19?& Baker, Gill, and SoIovay [3] initiated a study of rekivized complexity classes. Early studies

rewaled possliIe worIds in which P = NP = mNP, P # NP = uAW, or P # NP # dW hdds,

This appalIing phenomenon clear1y dec t s the ~ ~ c e in computationd power between determinism and

nondeterminiSlIt

In Chapter 5, we have aiready seen "reIativized computationn in terms of average-case versions of

Thing redua'bilitg. N a t u d y we can expand our boundary to r e b b i d wor1ds. This section wilt in-

traduce four relativized averapxm c o m p l ~ classes: Aver(P,F), Aver(BPP,F), Aver(NP,F), and

Aver(PSPACE, F).

The co~cept of-01~ is another way of viewing 'king reduciiility. w e have already seen two tgpes of

Thing redfta'bilitis: deterministic lhring reducibility and botmded-error probabIistic Thing redua'bility-

Here we shaU introduce the notation ~ v e r ( ~ , F ) ( ~ * ) to denote the coIIection of distniutional prob1ems which

are avpT-reducibIe to a given distributiod probIem (B,v), andogom to the nobtion ia worst-case

comp1- theorp-

Now Iet us iatrodttce a r e k b k t h of the fnndamental average classes Aver(P,F) and Aver(BPP, 7)-

w o n 63.1 (pelatmzea . - Aw(P,F)) [Sfl Let (B,u) be a dis&ibtrtid decision pmbIem- For

CHAPTER 6. AVERAGE CASE HlERARCHIES 206

a set F of distriiutions, denote by ~ v e r ( P , F ) ( ~ # ) the coilection of all distriiutiond decision problems

(A,p) such that there exist a deterministic o d e Turing machine M and a semi-distribution r )

the foIIowing conditions:

(i) (ESciency) M with orade B is polynomia-time bounded on paverage;

(ii) (V'dity) A = L(M, B); and

( i ) (Domination) p 5- q and t 2 Xt.@({z I z E Q(M, BJ))).

To improve readability, we simply say that the orade machine M computes A ioith om& (B,u) in

polynomial time on p-memge if Conditions (i)-(6) are witnessed by M w i t h some ddistriiution q.

We remark that it is posnile to introduce a weaker type of relativization using pT-reductions (see,

cg., [971). However, we believe that average po1ynomial-time Turing reductions are a better choice when

discussing average-case compIexity, becanse the dass Aver(P, *) is dosed under avpT-reductions.

P~~ 6.5.2 [97] Let(A,p) rmd(B,u) bemrydrPffibuftoncrl * . . decision problems, and let 7 be mry

set of d i d d & h s .

Proof. (I) Consider the oracle Thing machine that queries the mpttt string to oracle and then accepts

it exactly when it is in the o d e . (2) This is another interpretation of the t '-' itp property of avp-T-

reducibility. a

Fmally we s h d extend DeMtion 63.1 fiom a single oracle problem to a cJass of oracle problems.

Ropaitions 5 X 1 and 5-53 immediately yield the f b b w i q closure pro@es-

-6. AVERAGE CASEHIERQRCEUB 208

Next we show a colIapse of Aver(BPP,F) relative to Aver(BPP, *) down to Am(BPP,F).

prpposition 6.9.8 A ~ ~ ~ ( B P P , T ) * ~ ~ ~ ~ . ' ) = Aver(BPP, 7) for any set 3 of distribtrtions.

Proof. To prove this proposition, we assume that ( 4 p ) is in AV~~(BPP,F)(~") for some (B, u) E

Aver(BPP, *). In what fobws, we shaIl prove that (A, p) belongs to Aver(BPP, F).

Let M be a bounded-error probabilistic 'Ihsing macbiue which recognizes A with oracle (B, v) in polp*

mia time on paverage. We can amme without Ioss of generality that the error pmbabiliw of the machine

M wi th orade B on input z is bounded above by 1/4.

Let B' = {zOlk I z E B ) and Id

Note that (B'J) is prn-reducible to (B, u). Smce Aver(BPP, *) is closed under gm-redttaions, (B, 4 is in Aver(BPP, *). Using Corollary 3.532, let M,, be a bounded-error pmbabi ic 7 M ~ g machine which

recognizes B' in pofynomial time on d-average with the em>r probability R ~ [ M ~ ~ , ( z o ~ ~ ; s) # @(z0lk)] 5 rk.

Let us define the randomized Turing machine Mo as follows:

begin randomized algorithm for Mo w i t h an orade

input 2

simulate M on input x until the first query is made

if there is no qaerp then output M(z) and batt

f o r k = l to mdo

resume the simulation of M on input z

until the next query is made

(4 if M queries r then query dlkC3 instead and

receive an answer h r n o d e

if M fearhes a hating configmation then output M(z) and halt

encEfm

end.

Note that the pmbabiIity R , [ M ~ (z; s) = A(%)] Is equal to the probability Rah [MB(z; st) = A(z)j, which

is is at least 114. Thus, it is easy to show that A is rea@zed by Mo wi th orade (B',u') by an argument

similar to that for Lemma 5 . 5 2

Next we change the line (*) h m making a queq dlM3 to simnlating M,, on input dlH3. Let N be

the randomized 'lhring mruhinR obtained by this modification- By Lemma 5.53, it follows that the nmning

timeofthemache N is palyDomialonp-average

Consider the ermr probabiZay eN(z) of the machine N (MI input z. On each comprttaticm path generated

by Mo with oacle B' on input z, the error probability that N draws an emrnecms concIttsion is at most the

smn of all error probabilities which are made by the machine Mmp, whieh does not exceed=

Hence, eN(z ) 5 3/8- Therefore, (A, p) is in Aver(BPP, a.

This section introduces a relativization of Aver(NP, F) which win be used to b d d an LCaverage po1ynomial-

time hierarchy" in Section 6.4.

Schder and Yamalcami [9q &st studied an a- version of nondetaministc Turing reducibiity

and introduced a dativized class ~ v e r ( N P , F ) ( ~ p ) reqniring that dl computation paths be pruned to

the same wing a model of docked nondekmb&ic oracle ntring machines. Based on this model,

they introduced a " b i domination condition in such a way that it puts equal weight on all accepting

computation paths if one exists* or eke puts equal weight on all rejecting computation paths. Our model

of nondeterministic ' h k g machines is more general, and we cannot tak the same approach towards the

reIativizatioon of Aver(NP, 3). How can we define a relativization of Aver(NP,F) ? Steve Cook is credited

with the foIlowing idea

Let us reed the model of nondeterministic Thing machines. Our dehition of the nmning time of

nondeterministic lhring xuachines depends only on a shortest accepting computation path whenever it exists.

When we look at a computation tree of a nondeterministic l b h g machine on a particuIar input, we are

interested only m one shortest accepting computation path, pmfOvlded that one exists- Our domination

condition for the oracle machine needs a constraint only on the computation paths of interest.

Let M be a nondderministic o d e M g machine and let A be an o d e Let us recall that Acc(M, -4, z)

(Rej(M, A, z), resp.) denotes the set of (codes of) all accepting (rejecting* resp.) computation paths of M with oracle A on input z. We introduce a %p0opn set Fbp(M, A, z) as fo110ws:

By c&(z), we denote the (mde of the) minimal computation path of M with oracle A on inpat z if

one exists, and otherwise, the (code d the) minimal rejecting computation path. Obviody c&z) is the

minbd cornpatation path in FIip(M, A,z)-

Dehition6.3.9 (Rekhmd . . Aver(NP?F)) Let (B, v) be a dkki'bntiod decision problem- A dis-

tributional probIem (&p) is in ~ v e r @ P , F ) ( ~ p ) if there exist a nondeterministic o d e l h h g mae6ioe

Maudasemi-dIstributicmqsachthat

(i) (EEaency) M with oracle B is polynomial-time bounded on paverage;

(ii) (Validity) A = L(M, B); and

As in 'lhring reduciiiIi& we call condition (i) the d m a h t h cm&h fw M. WebeginwithatecbnicaI1emma.

ROOT. Assume that (D, p) Aver(lW, 7) (Ep) via a nondetermiaistic orade Turing machine M. There exist a semi-distxiidion 7 and s polynomial p~ such that D = L(M,E), p Wp rl, k . T i i e ~ ( x ) is p~ on

paverage, and ti(%) 2 @({x I z E Q(M, E, 2, &(x))}) for afl z. Without loss of generality, we a r e e that

~ m e E ( z ) > izl for dI 2.

Choose a polynomial p~ such that g is p~ on paperage. Moreover, let p be a p d y n o d and q a function

such that q is p on pa- and q(z) +(z) 2 P(x) for dl z. We can aPsome that an polynomiak, p ~ , p ~

and p, are mcreashg.

Now define a po1ynomial s as

where ~0 = h(A).

We shall show that h is s on patrerage For simplidtp, let A, and R, denote Acc(M, E,x) and

Rej(M, E,x), m y - Also Iet Ql, denote Q(M, E,z, y). Let B, = {z E CC [ Ti~nef&) 5

p~(1x1- 3r) A q(x) 5 p(1zl- 3r))- For any red number r > 0,

Cleadp the first two tams are bounded above by I/*- To complete the prooc we &odd show that the kit

term (q, Tr) is is asodolmded by I/*. Fix z E B, and asstme that h(z) > ~((21 r).

CIIAPTER 6. AVERAGE C A S E ~ C H I E S 211

Note that, for any set A, t({z I z E A)) 2 +((z [ 3 E Q,,afi(,)[z E A])). Hence, the bound on T, is

calcuIated further as fonow~:

Hence, we obtain the ineqaaity b({z I h(x) > s(r. 1x1))) < l/r, and this implies that h is s on paverage

0

The fo11owing is a list of basic properties which A v e r w , F)lBp) satisfies.

P r o p o ~ 6.5.11 Let (A,p) and (B,v) be d k t d d o d decision problems, and let 7 be cmy set of

Pmof. (1)-(2) Clear from the defh&ms. (3) The idea of the proof is Itimilslt to that of Proposition 5-5-7(5)- Assume that (At, pl) 9 (A2, p2)

via a reduction f such that (f,p~) E Aver(FP,*)- Asmme that a nondeterminiseic Thing machine M amp* A2 with o d e (A3,~(3) in time p~ on ~rraverage, where p~ is a polynomiaL Let MI be a

ClMTER 6. AVERAGE CASEHIERARCHIES 212

string z. Notice that

Next we consider the most conservative algorithm N which computes A1 using oracle A3: namely, on

input x, sirnuhe Mf (z) k t and then simnlate M nondeterministicaily on input f (z). Cleariy

for some pcsitive constaat c independent of z-

To simp@ the description, denote by Q, the set (w 1 z E Q(M, &,w, a(w))}- Then,

N m 6x w in Q= n ran(f)- For this w,

i i r (b I f (z )=~) ) h ( 2 ) - fit (4 PM (w) " z f (r)=w CM(W) ' ~f (21 - zg- &df@)~ * ~ f b $

To complete the p- we shodd show that ~z~Tiime$ (z) aud p are plynomid on pt-average since, by

Lemma 5-24 k-Tiie*(f(z)) tmns ont to be poIynomid on pz-aoerage. Eowever, by bp 3Al3 and

33-12, we can conclude that k-~he$(z) is poIynomial on ppaverage. S i the case for p is shdar, we

omit its p d

(4) To show the claim, we assmne that (Ap) E Di9t(P,*)(B+), and take an a r b ' i distniutional

problem ( D , a in Aver(NP,F)(b). There exis& a nondetc ' ' ' t oracle 'Ihring macthe MD which

cornpates D with oracle (Ep) in polynomiaI time on p-aveqe. We sMI show that (D,Q to

~ver(NP, *)(Bg)- Simce (Ap) E A~ee(P,*)(~p), we can find a demmb&ic orade Thing machine MA which computes A with orade (B,v) in polpnomia time on paverage.

CHrLPTER 6. AVERAGE CASE EIERARCHZES 213

Now we consider a machine M which nondeterministicaily simulates the compdation of MD, and whezr-

ever MD makes a very z, M determiaistically simulates MA on input z. By definition, M computes D with

the help of oracle B, and the nmning time of M with orade B on input z is bounded above by

where e is a mustant. Note that h.Tiie&,(z) is poIynomial on (-average, and the function

k- *EFWMD A) Ct EQ( M~ ,.E~,~) ~ i e ~ ~ ( z ) is also polynomial on c-amage by Lemma 63.10. As a result, Xz.TiieM(x) is polynomial on &average.

The proof of the domination condition for M is similar to that of Proposition 52.11(4). This yields the

desired co~l~equence that (D, Q E Aver(NP, F)(**). 0

W e c a ~ conjectnre that Aver(P,P-comp) M e r s &om Aver(NP,Pamp). As yet it remains an open

question whether this is the case in the t n d a h m d . * world In some rebtivized world, however, we can see

a clear distinction between Aver(P, P-comp) and Aver(NP, P-comp). Here we o h a relativized worId

inwhichthetwodassesdiffer.

Thetnem 6.3.12 T7me aists a problem (B , j r~) such that

ProoE We s6aII me the d e set constructed by Baker, Gill, and Solmy [3] to separate P b m NP. In the fdlowing prod, we first review their construction and then define the desired distri'butiona1 decision

problem (39 PB). AIso Iet {Mi)& be an U v e enumeration of all polynomiaEtime deterministic oracle l b h g machines.

Let a h &Ina be an enumeration of polynomials snch that each p,, safSes ~~rne:~ (2) 5 p,Jlrl) for any

choice of oracle 0 and any input 2. In parti&, the number of query strings of Mn on input z with o d e

0 is a h bounded abm by p,,(lzl).

Let us define the shialy increasing function C from N to N as fbIIows: let t{-I) = 0 for the sake of

convePience, and Iet C(n) be the minimal integer k such that k > C(n - I) a d h ( k ] < P. It is easy to see

that such s k exists for ang n.

We use the test langttage T(B) d&ed as T(B) = {On f 3y E Ply E B]). Tlis T(B) is a tally set and

beIongs to e. We shall construct a set B in the following snch that T(B) 4 PB. Fast we construct a

series of bite sets, {&)aaPI, and then let B be the tmim of all sets Bn- N m let B-1 = 0 for the sake of

commience. For each n E N, Iet yn be the minimll shing (in the standard order on C) such that

Such a string exists because IIQ(Mn,&-loL('))U 5 p,,(C(n)) < P(") by the d-on of l(n). Then, B, is

This shows that T(B) # L(Mn,B) for alZ nnmbers n, and thus T(B) 4 PB.

We then define the distribution p~ for the set B. Let

Consider the distriiutional decision problem (T(B), vay)- Clearly (T(B) , uWy) is in .4ver(~P, Pamp)(B+n).

Now assume to the contrary that (T(B), urJly) beIongs to Aver(P, P-c~mp)(~?"~) . There exists a determin-

istic Thing machine M which computes T(B) with oracle ( B , ~ B ) m poIynomiaI time on ut~lr-average.

Recall that T(B) is a tally set. Smce M is polynomial-time bounded on uuy-average, we condude that

T(B) E PB. This contradicts the fact that T(B) 6 pB.

F i i y we extend Dehition 6.3.1 h r n a single oracle problem to a dass of o d e problems.

Dehition 6.3.13 (Retativiaation) Let C be a elass of distriiutiond decision problems and let f be

a set of distniations. kt A Y ~ ~ ( N P , ~ ) ~ denote the rmion of ~ver(NP, atE*) for any oracle (E, v) c h m

h m C.

As an immediate consequence of Propositiort 63.11, the relativized Aver(NP, *) relative to Aver(P, *) to Aver(NP, *). Note that whether Aver(NP, *) relative to A=(NP, *) coJlapses to A=(NP, *)

is an open question.

Proposition 63.14 will be extended to any l e d of the average polynomial-time hierarcky in Section 6.4.

Another important average compkity class is the coIIection of deterministic average polynomial-space com-

putable sets, Aver(PSPACE,F)- This dasP contains Aver(P,F) and Aver(BPP,F) as subclasses- In this

~-an,weshauintrodnce - . -

a &hmatam of Aver(PSPACE, 7).

CHAPTER 6. AVERAGE C A S E l i E M R ~ 215

Deanition 63.15 (Bektivized Aller(PSPACE, 3)) Let (B, v) be a distributional decision problem.

A distn'butiod problem ( 4 p ) is in A~(PsPAcE,F)(~*) if there exist a deterministic o d e Timing

machine M and a semi-distniution q such that

(i) (EEaency) M wi th orade B is polynomial-space bounded on paverage;

(ii) (Validity) A = L(M, B); and

(ii) (Domination) p saw rl, and P 2 Xt j ( {z 1 z r Q(M,B,x))).

We also call the condition (iii) the dcmhatiun condition for M.

Since oar relativization is similar to that of Aver(P, F), the foIIowing proposition is straightforward.

Propo- 6.3.1 6 Let (A, p) and (B, v) & dist~+buthd decirion problenzs.

Roof. For the proof, we need a relativized d o n of the randomized bounded halting problem

(BHP1(A), p ~ ~ , ) . Assume that {Milia is an &kcthe enumeration of dl deterministic orade Tltring machines. For a set A, let

BEPl(A) = {(O,a,z, 1") I Mi wi th o d e A accepts z ming less than n squares )

~ { ( l , s ~ , x , P ) I Mi with o d e A on t uses at least n squares };

and, for b E {O, I), let

Notice that the distniation m~p, is not dependent on o d e A.

T& the desired set A so that A = BH&(A). We remark that this set A exists became Mi cannot qpery

any strings of length mare than n, and thus, it makes only queries to orade A that are lexicographidy

@er than (b, r, z, In). Now let M = mt. W e shall show that sny distributional problem (D, p)

in ~vet(PSP~c~,~cornp)(AP*) belongs to ~ver(P,~-com~)("+~) . For this pair (D,p), there exists a

d-c 'Ihring machine M c o m p ~ D with o d e A in polynomial time on pmmagc

Let g be the fimction d Lemma 4.2?(2) such that ji(x) < 2-m)l+2 f'or aJI z, and d&e the machine

Mt as fbIIows= on input z, it computes g(z) and then simaktes M on input g(z). Let i be an index such

~&=MrsinceMrisalsodetenninisti~Letns~nsikthef~deterministic~N:

begindetermiaisticalgorithm N input z

compute g" (2) and set u r g"(z)

forn=ltooodo

query (0, si, u, In) to o d e A

if (0, si, u, 1") E A then accept and halt

query (1, si,u, In) to d e A

if (1, sip U, 1") $ A then reject and halt end-for

end.

Notice that the macbine N makes queries only of the form (b, si, g" (z), In), where b f {O, 1) and 1 5 n 5 Tnne& (2) + 1.

It is not difiicult to check that, with oracle A, N computes D correctly. We next check the domination

condition for N. Remember that, hr each query string (b, q,u, In), the string g(u) is the only input on

which N queries it. Smw g is oneone, this correspondence tiom query strings to inputs is also onmne.

Thus, the rest of the prod is analogous to the proof of CIaim 11 in Section 5.3.5.

Let q(z) = 512- (lg(z)l+ 1)" .(Time& (z) + l)*/V-(si). It is obvious that q is plynomid on pawrage

because k.~ime&(z) is polynomial on paverage. As m Theorem 53.2, we have

This indicates that (D, p) belongs to Avet(P, P-cou~p)(~J')- a

We then introduce a relativized dass A ~ ~ ~ ( P s P A C E , F ) ~ for a dass C of distributionat decision prob

Ierns*

Deflnitioa 6.3.18 (Relatkkation) Let C be a cIafs of distriiutional dedsion problems. For a set

7 of distnIbutions, Aver(PSPACE, nC (Aver(PSPACE, F) dative to C) denotes the congaion of aII

distributiod problems m A~~~(PSPACE, T)(Bp) for some (B, v) E C.

PmoE It is clear that Aver(PSPACE,F') C A ~ ~ ~ ( P S P A C E , F ) * ~ ~ * ~ ~ ) . We shall show the

other inclusion. Assume tiat (A, Cc) E A~~~(PSPACE, for some (B, v) E Aver(PSPACE, *). Let

M be a deterministic o d e Thrbg machine which computes A with o d e (B, v) using poiynomial space

on paverage. Since (J3,v) E AverlpspACE,*), there exists a deta ' ' tic Turiug machine MB which

computes 3 using polynomial space on v-average The algorithm we would like to consider here is the

conserrrzrtive one: on input x, simulate M on z except for orade queries; instead of querying a string z,

simnlate MB on z. On each input x, this algorithm uses space at most

for same 6xed constant c > 0.

We can conclude that this bound is poIynomial on p-average since k ~ m a x , ~ ~ ( ~ ~ , , ~pa~eE(z) is poly-

nomial on paverage as in Lemma 6.3.10- Therefore, (A ,p) b d o w to Awr(PSPACE,F). 0

6.4 Average PoIynomiaI-Time Hierarchy

This section will f d p introduce an average-case version of the polynomiaEtime hierarchy. In previous

sections, we introduced relativized Aver(P,F) and cekmed - - A m p , F). These relativized cIasses wiII

be the basis for constructing an averagecase version of the polynomial-time hietiuchy.

6.4.1 Average Polynomial Time Hierarchy

We now give a fimnal dehition of an ctteruge polgmmial-Eime hiawrlry under a particular set of distniu-

tiom. This hierarchy is an averagecase d o p e ofthe polynomial-time hierarchy in worst-case comp1exity

theory.

-n 6.4.1 (Average PoIynomiaETii Hierarchy) [9q Let k > 1 and let 7 be a set of

distnintions.

Prod (I) In the case k = 1, the c[aim for Aver(NP,F) bIIm from Proposition 51.6. Now let

k 2 2 and assmne that ( A p ) 5 7 (3, v) and (B, Y ) E Aver(E!,F). By dehitbn, there exists a probIem

(C&) E Aver(Xi-,, *) such that (B, v) E A v e r ( ~ ~ , F ) ( ~ k ) . Proposition 63.11(3) implies that (A, p) E

~ver(NP, F)(Cd). Using the dehitim again, we obtain (A, p) E Awr(XE, 7).

For the cIass Aver(IIi,F), assume that (A,p) (B, Y ) and also (B, v) is in ~verm;, 7). Note that

(A,p) L)? (B, v) if and only if (zvCr) (B7 v). Ftom this fact, it h[Im that (x,p) E Aver(E!,T).

This is equident to (A, p) E Aver(IIf;.F). (2) SimiIar to (1). If k = I, then the claim is based on Proposition 52.6. For the other case, we use

Reposition 5.5.7(5). 0

In what fiows, we shan see the basic mclnsions among the classes of tbe average polynomial-time

hierarchy. Fttst we want to see the lemma that charaaerizes the relationship between Dist(E;,F) and

A=(CJ;, 7)-

h o E We shaIl show d y case (2) because case (I) foIlows by a similar argument. Let us assume that

(A, p) is in Dist(Chl, 7)- S i A E Z&,, these exists a nondeterministic made 'Ihring machine M which

computes A in pdynomiai time with oracle B in Xi. By a simpIe modikation of M and B, we can assume

the foIIorrPing proper@ on each input z, the madhe makes a query once of the form (z, y), where y is the

(ade of the) computation path. Note that, for every query string, there is the nnique pair of an mpnt and

apath, We shall show that (A, p) E Aver(NP, for some v, Let ns &st define such a distnIbutior~ Let v

Onr d* of v obvioasy gnarantees the domination condition for M- Hence, M recognizes A with or&

B in polynomial time Since (B, v) E Dist(Cg,*), we obtain:

CHAPTER 6. AVERAGE CASEHlERARCHIES

Using Lemma 6.43, we can prove the foUowing proposition.

Proof. (1)-(2) The claims foflow from Lemma 6.4.3. (3) This daim foIlows h m Definition 6.4.1 and from

the facts that ~ v e r ( ~ , F ) ~ A ~ ~ ~ ( N P , F ) ~ and (D,p) E A V ~ ~ ( P , F ) ( ~ & ) . 0

The above proposhion may be taken as evidence that our average polynomial-time hierarchy has a

structure s h k to that of the worst-case polynomial-time hierarchy-

Figure 6.1 illastrates the structure of the average poIynomial-time hkarchy under F-

Proof. Assmne that ( 4 p ) is in ~ v e r ( N P , F ) ~ ~ ~ ~ l * ' ) . By definition, we then have a chain of

membedip rdations (A, p) f Aver(NP, F)(Bp) and (B, v) E Aver(P, *)(cd), where (C, Q E Aver(Ci, *).

By Proposition 63.1 l(4), we can shorten this chain to (A, p) E ~ v e r ( ~ ~ , F)(Cx). This shows that (A, p) is

in Aver(C&, ,a. The converse is even simpIer to prove. Smce Aver(Ci, *) E Aver(AR1, *), it fo11ows that

Theorem 3.5.24 shows the separation between Aver(P, *) and AverlNp, *). Mthough these basic average-

case complexity clases are difhat, we do not conclude that s h k separations occur m the higher I d of the average polywmial-time hierarchy. The reason is that domination conditions restrict the complexiW

of-espedaIl~-- The woftt-case poiJmomial-time hietarcbp eutds the "downwatd maapse" propertg; that is, if any two

I d of the hierarchy colIapse, then the upper levels cohpe down to those 1eveIs. More precisely? if

EE = ELl, then PH = Et. Now we show that the assumption Aver@;, *) = Aver(CRt, *) Ieads to the

cow of the average pdymmiaI-time hierarchy.

Fgure 6.1: The average poIynomid-time hierarchy under 3

Proof. (1) Let us assume that .4ver(XL,, *) coIlapses to Aver(Ci, *). We want to show by induction on

integer i 2 k that Aver(CR, F ) Aver(Er,F) for all sets 3 of distriiutions.

The base case i = k is hivid b m our assumption- For the induction step i > 5, we have

Proof. Assume that (A, PA) and (B, p ~ ) are in Aver(C, F). The proof is by induction on k 2 I. For the base case k = 1, the daim for C E {P,NP) is immediate from Lanma 3.5.16. Now let us consider

the case C = ahw. Assume that (4 PA) and (B, pB) are hL Aver(mNP, 7). h other words, (&A)

and (Etm) areinAver(NP,F). AgahbgLanma3.5.16, wehave(hE,pA$pP) E Aver(NP,F). - Notice that a $ B = A @ B - {A}, where A is the empty string. Smce PA $ pB(X) = 0, we may ignore A,

and thm ( A t P A 8 p ~ ) E Am(NPtF)- This yields the desired condusion that ( A $ B, @ ps) E

Aver(cpNP, n* For the induction step k > 1, first let C = X!. By our assumption, there exist distn'butional problems

( C O , ~ ) and (C1, &), both of which belongs to Aver(PL1, F), such that (At pA) E A=(NP, F ) ( C o A )

sod (B,CB) Aver (NP,F) (C1~) - It ig &tidy to ~ e e tbat (Co,&) sk (Co @Cl7Co @PI) a d (el,&) 55 (Co@Cl,Q@pl). Forsimplicity,write (C,c) for (Ci$Clt[o~pl). By ourinduction hypothesis,

we obtain (C, Q E Aver(Ek,, 3'). Using Lemma 7212, we get ( A d By M @ pB) ~ v e r ( ~ ~ , T)(C*). As (C, Q is in Aver(E:-,,aF), the pmb1em (A $ B,pA LA p ~ ) is in Aver@';, F). The case C = A: is similar?

and theother a s e C =IIL isobtained withthesameideansedfor thebasemseC=c+NP. a

At last, we may conjecture that, for a naturally selected set 3 of disttz'butions, the average polynomid-

time hierarchy tmder 3 is truly an infinitr?. hierarchy.

6.4.2 Sparse Interpolation Property

We shnn show a basic relationship between worst-case c o m p l e and average-case complexity on strings

w i t h high probability. We kst introduce an interpolction property of an amagxme compIexi@ class

Aver(C,F)- IntnitiveIy, the property says that if we cornpate a set A fast on average tmder a distriiution

which assigns high pmbabllity to instances in a sparse set S, then there is an "interpolad" set B between

A n S and A which is computabIe fist in worst-casec This set B is s coIIection of "easyn instances m A, and

it becomes a good worst-case approodmation of the set A.

Here is the f o d dehition of the sparse ~~ propertyproperty

Deanition 6-4.8 (Sparse fnterpolation Roperty) [9q For a sparse set S and a polynomiaI q, let

fbrdlx~SptOVidedthat)=,~&~-(l. AcIassAver(C,F) h a s t h e s p o r ~ e ~ ~ ~ i & f o r

any set A, any h h i t e sparse set S, and any polynomial q su& that (A, pS,J E A-C, F), there exists a

set B E C suchthat A n S c B S R Theset Biscalledan interpohtofAasdS-

P r o p o ~ 6.4.9 For a doss C E {P, NP. RP, BPP, PSPACE), Aver(C, r) has the sparse interpoldm

-*

Root We first show the case C = NP. T& any sparse set S and a po1ynomial q and assume that

( A ~ s , ~ ) c Aver(NP, *). There exists a nondetermiaistic Turing machine hi which computes A sach that

b - m m e ~ ( z ) is p on psPq-average for some polynomial p. Note that Tune&) 5 p ( l ~ l / j i ~ , ~ ( x ) ) for an x with Ps&) > 0. Let N simdate bf on input x in p([xl q(lz1)) steps. If the simtllation of M does not terminate

withiu p(lxl- q(lz1)) steps, then N rejects z- Let B = L(N). Clearly B A. S i q(lzl) 2 l/fiS,,(z) for dl z E S, N compIetely simulates M on all inputs z in S. Thus, A n S = B n S. Clearly N is polynomia-time

bounded. Therefore, B f NP. The other ~ s e s are treated simhrly, but spedficdy for the case C E {RP, BPP), we must use Tmeb (z)

instead of (z; s) . 0

Proposition 6-42 can be extended to an arbitrary I d of the average po1ynod-time hierarchy. However,

its proof is not as h p I e as that of Proposition 6.4.9. W e first present a key 1emma.

Proof. (I) The proof proceeds by induction on k. The base case k = 1 essentiaIly fotIm h m Proposition

6-49. Assume that (A+) E A m , * ) and Iet S C C and let q be any poIynomid_ There are a

nondeterministic ntring marhine M and a polynomial p such that A = L(M) and B({z [ TiieM(z) > p([z[ - r)}) < l/r for d r > 0.

Let r = q(n), and then b({z 1 Ti&(z) > ~(1x1 q(n))}) < l/q(n). Let Mo simnkte M on the same

input in h e p(n -q(n)); if M accepts z, then & accepts it, or eIse Mo rejects z. SimkIy, let i& simalate

M in time p(n - q(n)); if M rejects z, then Ml accepts it, or else MI ~jects it. Define Co = L(Mo) and

Cl = L(Ml), and let Sf = Sn (G uC1). Clearly A n St E C' E A and an Sf c Cl H Momovert

The mdaccion step is carried out as fbIIows. Let k 2 2 and amme that (A+) E Aver(cE, *). By

definition, there exists a distn'brrh'onal problem (B, v) E Aver@:-, , *) such that (A+) € A-, F)@P). Let M be a nondeterministic o d e l b h g machine which computes A with oracle (B,v) m poiynomial

time on p-aveqp We armme that ~ z ~ T i i e ~ ( z ) is p on pavetage fix some inmasing po1ynomial p.

The domination condition for M impIIes the exisfence of a semk&ri%ution q and a fimction d which is

CaAPTER 6- AVERAGE CASE EiERARCHIES 223

pdyttomial on paverage such that d(x) . ij(x) 2 fi(x) and B(z) 2 ij({z I z E Q(M, B, z, &[z)))) for aU z and z. Assume also that t is a polynomid witnessing that d is polynomial on pawrage-

Consider any set S and any poiynomial q. We d&e T as

For each n E N, let us consider the subset T" of T. Notice that

In the rest of the proof, we use s(n) for t(n - 3q(n)). To use the induction hypothesis, we let Z = { z I 3z E T(z € Q(M, B, z, &x)) ) ) , and we also let

Applying the induction hypothesis to Z and I(n), it foUows that there exist sets 2' G 2 , Ci E EL,, and

C'i E I& such that, for every n E N, P ( P ) - Q(Zm) 5 I/l(n).

Now recall that M queries strings whose Iength is at least the same as that ofthe input and at most the size

ofits nmningtime. Inother words, foreveryz mTn, ifr E Q ( M , B , Z , ~ ~ ( X ) ) , then n< Izl ~ p ( n - 3 q ( n ) ) .

Based on this Fact, we prepare sets &, = {z E Z I n 5 Irl 5 p(n 3q(n))) and 2; = {t E 2' [ n 5 121 5

p(n - 3q(n))l- sme ~ n - ~ : r ( ~ - 2 3 n { ~ ~ ~ ~ l z l s ~ ( n - 3 4 ( n ) ) ) ,

it foUows that

The desired set Sf is d&ed as Mows

Clear1 y Sf S. Using the fact that

by the domination condition, rn obtain &(gn - 23 2 ij(P - Sm). Then,

b(W - fi(T=) < i;i;;r. we et(W - b ( P ) < A- kMokanarade~~withoradeX~~foII~~~.Oninpatx,M~sirmrlategMm

z m time p(]d -a([@), and whenever M queries a string z, No queries both Oz and Lt to its o d e X. If

CHAPTER 6- AVERAGE CASEHIERARCHIES 224

Oz f X and lz fZ X, & continues the simulation wi th the assumption that the made answer is "yes"; if

Or 4 X and lz E X, then it continues the simulation with the oracle answer "non; otherwise, it immediately

rejects the mput 2. The machine MO accepts z exactly when M halts and accepts it. Similarly, we dehe

a machine MI by interchanging the oracle answers and requiring that MI accept the mput z if M halts in

time p(Iz1 3&[)) and otherwise MI rejects z. Now let CO = L(Mo, Ck $ Ci) and Cl = L(M1,C; $ Ci)-

By thedefinitionsofthecuaclemcbes MOandMl,itfolIowsthat AnSrCCo ~ ~ a n d h s ' c ~ ~ GZ. (2) The proof is similar to (1). n

Proof. We show only the case Aver(CE,*) here. The case k = 1 follow &om Propositim 6,423. Let

k 1 2 and assume that (-4,~s.~) E Aver(C,P,*) for a sparse set S and a polynomial q. It foIIows h m

Lemma 6.4.10 that there exist a set C E C! and a subset S' of S such that A n Sf C E A and

&S,~(S") - f i ~ , ~ ( S ~ ) I 1/2q(n) for all n E N It d c e s to show that Sf = S. Asnrme that there exists a

string z E S - Sf. Let n = IzI. Smce fi&) 1 l/q(lzl), it folIm that

This is a contradiction. Hence, Sf = S. 0

The following proposition is another consequence of Lemma 6.4.10, and it win be nsed in the next chapter.

Proof. Let p be any poIynomial A s m e (A, V-) E Aver(CE, *). Take q(n) = 2(n + I)' - p(n). Note

that

We apply Lemma 6.410 to q, A, and C". Then, we have sets Co E CI, 4 E Ti:, and S such that

A I ~ S S C O ~ 4 X f l S ' ~ C I SZ andv-(C-Sm) s l fq(n) f o r a l l n ~ n ObvionsIy,SS'CCoUC1.

Hence, IIq U qn 2 IIS"I[. It scrtlices to show that [ISfnl[2 (1 - A) 2"- Note that

Smce kroad(P - Sn) < l/q(n),

6.5 Average Polynomial-Time Alternation Hierarchy

I . wombcase w m w theory7 alternation plays a varietJr of roles. R e d that AEf; = ATIMEE(k7 nO(l))

and Mf: = &(k,nO(')). Then, we have AEE = C: and Ah: = A: for all k > 0; namely,

the poIgnomiaI-time alternation hierarchy is d y the polynomial-time hierarchy (see Lemma 2-5.1). In what fobws, we intraduce an average version of this polynomial-time dternation hierarchyt an average

polynomia-time aIternation hierarchy under a set of distriiutions, and study its relationship to the average

poIynomial-time hierarchy de6ned in Section 6.4.

D-tion 6.5.1 (Average Polynomial Time Alternation Hierarchy uuder 7) Let k > 0 aud let F be a set of diski'btttio~~. The aoaage p o ~ - t i m e alternating hiermchy under 3 consists of the

fonowing average classes:

1. Aver(AAi, F ) = u ~ > ~ AVE~(ATIME~(~, nC f c), F).

RecaIl that I-alternation bounded alternating 'ihring &es with e x k h t i d states are exactly the

nondeterministic Thing machines. Similarly7 semideterministic l-alternation botmded dternating Thing

machines are determrmstlc * . . ntring ma&hes. Hence, it is dear that Aver(AA;,F) = Aver(P,F) and

Aver(A.X;, 7) = AverlNp,F) for all set F. NePeftheIess , two~Aver(AI:~ ,T) andAver(CE,F) are~ytobethesameeveu6orthesetPof

~ i I y computab1e distributions. The folhwing propasition heIps us tmderstand the gap between these two

ches. Recall that Aver(AR,, F) = ~verfP, F)*-:*') and Aver@Kl F) = .4ver(NP, F)*-:**)-

Prod We pwe (2) here. Fnst are assmne that (A+) E Aver(ACR,,F). There exists a (k + I)-

alternation bounded Thing machine M gnch that A = L(M) and Xz.Tii~(z) is p on pammge, where p

is a polynomial We shall show that A is computable by a nondeterminirtic orad= Taring 'Lhring N with

oracle ( B p ) i n ~ t i m e o n p a v e r a g e .

CEQPTER 6. AVERAGE CASE HIERARCHIES 226

The main idea here is to query sufficiently long strings to the oracle so that the complexiq- of the oracle

set can be substantially rliim;nished. We begin by deb@ the oracle Thing machine N as follows

begin nondeterministic algorithm for N

input z (assuming that z # A)

simdate M on input z during the first series of dikntkl states

until M reaches the universal states

(let y be the code of such a computation path)

choose a path sbrting hzn y nondeterminiSticaIly (ignoring the states of configurations)

and folIow this path until M reaches a halting configuration

(let n be the length of the path from the initial con&mtion)

query the pair (2, ylOn) to the d e

if oracle answers "yes'' then reject eIse accept

end.

Clearly N is a nondeterministic ntring machine whose nmning time is O(TiieM(z)).

Let 3 be the set that is accepted by the bllowing dgorithm. On input (z, ylOn), check if y encodes a

computation of the &st series of ex&ntbl states of M on input z, and if so, then sirdate this path tmtd

M goes into mriversal states. Sinlate the rest of the computation of M within n steps. If a path does not

hat within n steps, then go into an accepting co&gmation immediatdy. It is not di6icult to see that B is in

ATIME'(~, ~ ( n ) ) , and thus B E C:. By definition, A = L(N,B). To see that (A, p) E o n , 4 v e r ~ , F)@*3

for some u, we remark first that Q(N, z, &)) = ({2,y10n)) for some y and n. Next we define

It is easy to check that (4p) E AAP~~(NP,*)(~*) via N. !%hce (B,v) is in Dist(XE,*), we have (A,p) E

Aver(~~,7)==t(%').

Comedy, assmne that (A, p) E A=(NP, 7)h(4b'). There exists a problem (B, v) € Dist[E!, *)

snrh that (A, p) E her(NP, F)(B+') via an appropriate nondeterministic oracle Zhring machine M. A s m e

that B is in &(kYp(n)) for some poIynomiaI p- We can asnre that p is strictly increasing.

Next we define an dternating lhk tg marh;ne N as foIlows. On input z, N &mdates M on mpnt

z starting w i t h an ex&ntbl state. Dnring this akkntid state, if M queries r, then N stores z Cm a

work tape) and guesses its oracle answer m(t) and continue the simuIation until M terminates. On each

mmpatation path, if M reaches an accepting m@mation, then N existentially checks if s E B for all zi

withons(q) = 1. Then N nnioersally chedrsif~ $ B for aILq whichsati&m.s(z+) =Oat onceusing B.

If=(&) = 1 fix all such s, then accept 2; otherwise, q*ect z. It is easy to check that A = L(N). The nmning time af N on input z is at most

for some constant c > 0. Since Xz.TmeM(z) is polynomial on paverage, by Lemma 33.14, h . T i i e ~ ( z ) is

also polynomial on paverage. We then have (A, p) E AVIX(ATIME'(L + 1, no(')), 7) = Aver(AER1, 7).

0

Coro1hn-g 6.53 Let k > 0 and 3 be a sef of d&+h&m-

In worst-case c o m p l e theory, there is a nice characterization of the class PSPACE by polynomial-

time alternating Thing machines: that is, PSPACE = ATINE(no(')). Interestingly, WE & not h o w

whether Aver(PSPACE, F) equals ~ver(AITbfE(n~(~)),T) becanse of the definition of the running time of

alternating ' h h g machine.

In what foIlows, we s h d show that Aver(BPP,F) for a supportive set 3 of distriintions is Iocated

within the average polynomiaLtime alternation hierarchy under 3.

Proof. Asmme that 3 is supportive and Iet (D, p) be an arbitrary problem in Aver(BPP, 7). Smce p is

supportive, we can define a strictly increasing, positive polynomial p such that P(z) 3 2-p(I21) for an z. By

Proposition 3.533, there e x h a randomized Thiag machine M which recognizes D in po1ynomial time on

paverage with error probabilrty 2*1~1)~, ie., &[M(z) # D(z)] < 2-~(l"l)'.

Now Iet us d&e

Apply Lemma 3331 to Xz.TrmeH(z), and we conclude that Xz.Tmeb(z) is polynomial on paverage Let

q be a polynomial such that kc-Tiieb(z) is q on paverage. For this q, we have Timeb(z) q(fxl//i(z)) d e ~ ~ P(z) = 0. Let S = {Z I i (z ) > 0). Then, for a h s t all z in S, Tie;la(z) 5 q(lz1 -2P(lzI)) < 2P(lrl)2-2.

We take a positive integer satisfying Log(Tiieb(z)) < p(1z1)2 - 2 for all z E S wi th 121 2 no. We next cIairn that the pmbabm p, = Pr,[M(z; s) = D(z) I TimeM& s) I Tiieb(z) h s E R&)I

is at least 1 - 2p(12 I)'+'. ~rrst we

CHAPTER 6. AVERAGE CASERTERAR- 228

Notice by the definition of Ti ieb(z) that P r . ~ m e e ~ ( z ; s ) 5 Tiieb(z) I s E R&)] 2 1/2. Hence, the

conditional success pmbabEZity p, is at least

In particularl when m = TimeL(z), p, 2 1 - 2-~(l'l)~+l > - I - 2-Iqrn-l since logm < p([z1)2 - 2.

Let us d&e another Tkhg machine M' that sim&tes M by using an additional input string as a

random seed- Formallyt the algorithm is as f01lows:

begin det erministic algorithm for M'

inpd (w) (say y = ytyr * * - vm, where yi E {& 1),1 I i I m) let i -0

start the simtllation of M on mpat z

wbile the simulation do

if M fips a coin and i 5 m then let its outcome be yi

i f M flipsacomandi>m then enteran infiniteloop l e t i = i + l

end-while

end.

Note that we do not re- the machine M' to exhaust all bits of y on each computation path of M on z;

thus, M' halts on input (2, y) exady when s is a p& of y for some random seed s f ~ M ( x ) -

Towards achiwing our goal, it sufEces from hposition 6.52(2) to show that (D,p) belongs to

~ v e r ( N P ~ F)(E*) for some dbtri'blttional problem (E, v) m Dist(NP, 4. We first define a nondeterministic oracle lhring machine Mo as follows:

begin nondetermisistic algorithm Mo with an oracle

input z

if < no then output D(z)

guessastringto

let m = lwl

simulate M' on input (z, w) for rn steps

if either M' does not halt within rn steps or M' entets a rejecting state

then reject

guess m distinct strings ul, ...,% of length m

qpery the string ( z ,u~ -u,,,,w) to oracle ifthestringis notin the oraclethenaccept elsereject

end.

W e now show that, for any oracle, & is time bounded on p- To show this, it d c e s

to caasider the ease where b[ = Time&(z). F i notice that, fbr any onYe 0, TiieL(2) 5 c - (m +

C&APTER 6. AVERAGE C ' H z 3 B A . R - 229

m2 + I), where c is an appropriate constant. Since k . T r m e ~ ( z ) is pdynomial on p-average, the hct ion

k.Tiie&,(z) is also polynomial on p-average.

Next we shdl define the desired d e E so that D = L(Mo, E) . The set E is defined as the set computed

by the nondeterministic machine MI below. Let u @ a denote the bitaise addition of u and o moddo 2, and

let the algorithm for Ml be as follows

begin nondeterministic algorithm for ML

input (z , ul - - -u,,,, w)

if lut - - # Iw12 then accept

(Now assume that rn = Iwl-)

guess a string v of length m

for i=l tomdo

simtrl;rte M' on input ( z , ~ $ v ) for m steps

if M' does not halt then accept

simtrlate M' on input (2, w ) rn steps

if M' does not halt then accept

if Mr(z , ui @ a) # Mr(z, w) then accept else reject

end-for

end.

Note that MI is polynomial-time bounded, and as a d t , E belonp to NP. Still we must prove that

D = L(Mo, E). For sim~licitg.~ we fhc x and set rn = Treb(z). Let A = {(w, g) E P x X"' 1 Mt(x, w) = M'(z, y)}. Moreover, let

B = {w E C"L I there are more than 2m-1wm-L strings y of length m such that (w, y) E A}.

The cMtinalitJr0f the set Cjo thus is bounded above by

(If - part) Assume that the right side of the equivalence in the daim is true; namely, there exists a

m-tuple ( u ~ , . . ., u,,,) E (IF)"' such that, for every v f P, (w, $ v ) E A hoIds for some i. Fm such an

m-tuple (ul, ...,u,& Let C = {u € P I (w,u) E A). For each y E Ct, there exists a number i snc6 that

(wteu) € A . Foreachi,let C:= ( ~ E P I ~ ~ [ ~ = W $ U A ( W ~ ~ ) EA]). Becauseofthedehitionof~,

11411 = IlC'll for all pairs (i, j ) , 1 i, j 5 m. Since F = r=, C:, we have IICill 2 Zm/m > 2*-t0p-1.

Thdore, w is m B. II

Reedl that x E D if and only if there exist more than 2m-'q(m)-1 strings y of length m such that

( q g ) E A- This is equivalent to saying that there exists a string w of length m such that w E B and

M f ( z , w) = 1. By the above cIaim, it h o b that

This yields the equation D = L(Mo, E). Finany we d e h e the desired distniution v on E as foIlons. Using the frmction c&,,(z), we set P(z) =

i((z z E Q(M& E, z, 4, (z)) )). It is to see that (D, p) is is inAw(N, 3')(&~), and

co~l~eqlfentiy (D,p) is m ~ v e r ( N P , 7 ) ( ~ ~ ) ~ v e r ( N P , 7 ) ~ ~ * ' ) = Aver(AE!,F).

The other daim that Aver(BPP,F) c A~er(An,~ , f l f o l m h m the mdusiom:

6.6 Average Low Hierarchy

The atrerage poIynomiaI-time hierarchy ~ O W S as to construct an a- version of the toto ) r imdig

in NP to refine the stntcture within Atter(NP,F). Pethaps some NP-compIete probIems with nattxral

distri'butiorts which are unknown to be either m A m p , *) or p ~ m p I e t e for Dist(NP,P-comp) fail into

a Iorvt ldaardrg in Aver(NP,P-cc)mp)-

We dehe the relativized average poIyrtOmiaEtime hiefarchp-

Definition 6.6.1 (Relativized Average PoIynomiat Time Hierarchy) Let P be a set of distribu- tiom. For a distribntid decision probIem (Bp), the relaCioited wemge polynomial-time hkmdiy &

3 d d i v e to (B,u) amskts Ofthe fo-g classes

(2ECIER 6. AVERAGE CASE IZZZMU- 231

Dehition 6.63 (Average Low Hierarchy under F) Let k E N

Here we remark that it is open whether each dass of the aPPrage low hierarchy enjoys the sparse inter-

polation p r o w .

Hence, A w ( A L , , *)(**I E Aver(AL,, *), which implies that (D, p) Mow to ~ver(LAk,, n. Snnl't;rtly, we can prove the other indnsion. I3

The assnmptiiopl A m : , *) = Alter(c&,, *) is & a t fm the dao Aver(NP, 3) to coaapSe to the

kth level of the average !ow hieratcbg-

CHAPTER 6. AVERAGE CASEafERARCHES 232

Proof. Let us assume that Aver(Cg, *) = Aver(CK,, *). Let (D, p) be an arbitrary distributional problem

in Aver(NP, 7). We shaII show that (D, p) belongs to Aver(LCE, 7). Since (D, p) E Aver(NP, F), we have

the following inclusions:

Aver(Cg, *) E Aver(Zi, *)(D*) G Aver(CR,, *).

Unfortrmately, we do not know any natural w p l e s of distriintional decision problems falling in the

average low and high hierarchies. The search for such problems is a chanenge.

Chapter 7

Quint essent i d Computability

7.1 Introduction

The most exciting aspect of this thesis is the attempt to investigate the notion of q r t i n t e s s d cmp&MdgT . .

first proposed by Schaler and Yarn- (911. Throughout this thesisT we have developed the avaagecase

complexity theory initiated by Levin. We know that a- compIexity theory is very sensitive to the

choice of distziiutions. For example, if we take a distri'btttion which d m h t enough to 0, then all NP problems are polynomial time soIvabIe on the a- N d e l e s ~ , this type of extreme analysis does not

capture the signiscant feature of average-case complexity theory-

Regarding Levids -on Dist(NP,P-eomp) cAver(P, *), Ben-David, Chor, GoIdreich, and Luby [9]

gave a partially negative answer by dernonstratEng that Dist(NP, P-comp) Am(PT *) d e s s E = NEL This reslllt sttggests one approach to open westions posed m average-case compIexity theory-. embedding

amagewe ampi- dasses into worst-case comp1- theory in such a way that the embedding does

not lose the compIexity of these classes. The simpIest s01ntion is to focns on "rare instances" under (Ceverf'

reasonab1e diss l iat io~~ ThiP notion was developed by SchuIer and Y k d a m i [97j (suggested by Uwe

Sch6aing) . In Section 7.2, we shall formally introduce the notion of kd C under 7." For a (wok-case) mmp1-

class C, %d C mder F represents the class of sets which can be comprttaible on average no matter

what distxi'butions are chosen h m E The simple notation CF was invented by Schder and Yamakami

[ 9 to denote the dass 'real C under 3$ for exampIe, P h m p captures Yred P under P-eomp." Using

this notation, Levin's question Dist(NP, P-comp) c A e T *) can be simply rephrased by the question

NP E1pp-comP in the worst-case setting-

Based on this notion, we are able to iotroduce the d p o w - t i m e k t m h g mcda 7, Xic, qF I k > 0), that esptares the essartiab d the average polynomiaL-time hierarchy under 3. By the

deWtion, -y computabIe dasses ha;oe oery difkent structures from their worskase connter-

For instance, it is mhn,m whether Pr eqaals Uk,,,DTIME(O(nk))r m gemmi, whereas P is

CHsLPTER 7. QUINTESSENTIAL C 0 M P V T A . B ~ 234

Uk.,,, DTIME(O(&)). One of the exciting results here is that the real polynomial-time hierarchy under the

set of recursive distniutions indeed coinades with the worst-case polynomial-time hierarchy- In particu-

lar, Pbmp c05pse~ to P. These &ts support our main focns on feasible distributions in average-case

analysis. Section 7.2 wilI f o d y d&e the notion of "real C under 7" and demonstrate the fundamen-

tal propertks and the eqttivaence betwen the red polynomia-time hierarchy under REEcomp and the

polynomiaEtime hierarchy.

Our interests are now in the real polynomial-time hierarchy under P-comp and its alternation connterpart,

the red --the alternaPi4n IrierOrchy under Pcomp, and we shaU study its properties from the

perspeaive of structural behaviors. The first question we wad to ask ourselves is whether any level of

the ceaI poIynamial-time hierarchy under P-comp contains sets which are hard to compute. Schuler [92)

succeeded in constructing such a hard set within Phmp which cannot be computed by a deterministic

polynomial-time Thing machines. His method is further extended by Schder and Yamagmi [98] to the

separation between DTIME(0(2Cn)) and P-,. In Section 73.1, we sha3 see how to construct hard sets

within and Cg-,,, k E y by using resourcebounded KoImogomv complexity.

Inuarme sets and bi-imwme sets are good examples of hard sets. The dass E, for example, contains

P-immune sets and P-bi-imwme sets, and thus E is different h m P. We shall see that Pkmp contains a

P-immune set of an " a r b i i density. Nonetheless, the class P-,, or even its &tabIe dosure has

no P-biiimmune sets, and as a consequence, the class tmns out to be s d l , i-e, it has pmeasure 0, horn

measure theoretic point of view. This consequence is intriguing in contrast with the faa that the Thing

d o s a r e o f P ~ , , isequaItothedassEXP,whtcbhasprneasure1.

In 1974, Book [15] Erst showed that E is JtruetwrJly di&rem from NP. He actually proved that NP enjoys the cIosure property under pm-reductions, but E does not; therefore, NP cannot eqml E. A simnar

structural,.property characterizg the classes in the real polynomid-time h i d y under P-comp. We shall

see in Section 7.5 chat neither nor S k m p is c l o d under pm-rednctions, and consequently

both A;-, and CE-, are slncduraCIy di&rent from all worst-case complex& cbss which are

cIased under pm-reductions.

In Section 7.6, we shall look at the probabilistic dasses BPP-, and B P P M P 7 which are another

example of weIIaadied quhksmtial comp1exity classes- h e to Ira- and Levin [44 and W e r

and Watanabe [96], the question NF mPPp-com, is known to be equivalent to the question NP C_

? B P P h p . This is not howa for P-, and P-,. The notion of random oracles was introduced by Bennett and Gin [8] in 1981 to prove that the probability

of the event of a d a i i v k i NP coinciding with a relativized P is 0 when oracles are chosen at random.

In Section 7.7, we shall show that neither of the inclusions, NP G Pp-com, nor PPSomp 5 NP, is passiile

relative to a random o d e .

Major Contributions. The author fiwmaks the notion of snbstantiaI computation on the average and

h the hardness and the strnctnral properties of the classes m the real poiynamial-time hierarchy tmder

Pump-

Propositions 7 2 5 and 72.10 show the inclusions among qufntessential computable classes: e-g., A:

AAEF E A ~ F G At,. A similar inclusion holds for the XI; class.

Propaition 7114 shows that, for any dass C of the polynomial-time hierarchy, every set m tin is nnearIy-

C. As a direct consequence, Corollary 72.15 shows that if strong one-way functions exist, then NP is not

included in BPPF for any set of distriiutions 7 which contains at least one supportive distribution.

Lemma 72.16 shows that if Aver(C,F) has the sparse interpolation property, then TAUY n CF C C. This Iemma leads to several interesting consequences including, as m ComUary 72.19, 33: E AHr unless

A., = Pk.

Theorem 72-23 s h m that AEREC-COmp = AAEREC-eomp = A! for my k. S i I y for CiRM=-p and BPPREC-p- In particular, as m Corollary 72.24, the class PEwmp equals P.

Theorem 7.32 shows that, for each constant c > 0, there exists a spa,= set in Mi-, but not in

m A ( k , 0 ( 2 ~ ) ) . A similar qmat ion result holds for AC:-mp-

Proposition 7.3.4 shows that, assuming PPamp # P, either FP g #P or NP BPP hot&.

Proposition 7.3.6 shows that for each constant c > 0, MEkmp is not included m

e ( k , dose. This immediately implies that P-, P-dose.

Theorem 7.3.9 shows that, for each constant c > 0, Aft'-,, is not included in A,P/cn.

Theorem 7.41 shows that there exists a P-immune set m P-,. The final cIaim of Proposition 7.4.5,

that there is no P-bi-immune set m PhmP, comes fiom W e r and Yamakami [98].

Proposition 7.5.10 shows that Pimp and Xi-, are closed under hpm-reducti~ns.

Lemma 75.9 shows that there exists an incomparable pair of sets A and B in P-, with respect to

hpm-reducibility, namely, A &' B and B $2 A.

Proposition 7.5.8 shows that if every P-samplable distriiution is avp-domhated by some Pamputable

distriiution, then AEpTomp is closed downward tmder hpm-reductions.

A s for poIynomiaUy-bounded operatoft, Theorem 7.5.13 shows that the class 3AEp-, is not included

in A~h,p. A s its coroIIarg (CoroIIary 7.5.14), P-P-, N P k m p ifP = NP. Theorem 7.5.16

asserts a similar result about the probabilistic operator PP that the class PP. P-,, is not induded in

F+FccrrnP.

There is a series of random oracle separation results taken from SchuIer and Yamakami [97,98]. Propo-

sition 7.7.4 shows that, reIative to a random oracle, PmP is ddkmt h m NPWp. In Theofem 7.7.6,

it is shown that, dative to a random oracle, NP is not included m P w , , W ~ ~ L W S Themem 7.7.7 shows

that, relative to a random oracle, PMp is not included m PSPACE.

7.2 Red PoIynomiaI-Time Hierarchy

Average-case compIexity theory has given us a different perspective hm warst-case complsritp theory about

what is hard to compute. Some NP-compIete problems have been M y categorized as relativeIy "e&

on the average- This is one of the reasons that Levin asked whether all NP-compIete probIems are %a@

ontheaverage

As shown by Ben-David, Chor, GoIdreich, and Luby [9], Levin's question is closely rekted to the E =?NE

qttestion in worst-case c o m p l e theory. Is there any general way we can disctrss other open questions

m average-case complexity theory in tenns of worst-case compIexi@ cIasses ? Or more bravely* can we

embed om amqpase compIexity clase!~ mto the world of worst-case complexity cl;rsses without losing

atly si@cant feature oftheir average-casecomplexity ? One auswer was presented by Schder and Yamakami

[9q m 1992. They introduced a new notion, d e d d C under F, that enables us to treat an aoeragpease

complexity cIass as a worst-case complexity class.

This section win begin with the notion md C under F and then introduce the d plynumidtime

himmhy.

7.2.1 The Notion of 'CReal C under F

At a conference in 1992, Schtrler and Y M [91 proposed a way that we can bring amageae complex-

ity classes back to the worst-case complexity world. The idea is that rather than argue the average behavior

of an algorithm with respect to each individtta distniution, we wish to extract the hardest instaoces under

eoay distrifiution. Let us consider a set A m P. The set A is not only computable in po1ynomial timet but

also computable m p1ynomi;il time on pavetage under every distniation p In other words, A is ''easy" to

compute regardless of the probability with which each instance occurs. What kind of instances are to

compute on average nnda aU maonable distriiutions like P-computabIe distnLbuti011~ ? We shall formalize

the co11ection of such instances in a more general way.

Formally, we introduce the general notion of "rea C under 3."

De5ition 7.2.1 (Red C under F) [9l] Let C be a complexity dass and let 3 be a class of distniutions.

Assume that Aver(C,F) is defined. The class nol C unda F* symbolically CF, is the dass of sets D such

that (D, p) E Aver(C, *) for every p E 3-

This new notion f 0 n d . k ~ a significant fatme of the associated a- compIexity classes. The next proposition indicates its importan=

PtooE A s s m n e t h a t C ~ ~ s n d ( ~ p ) i s i n D i d ( C , ~ . brnthefaathatAbelongstoVFtitfo~m

that (A, p) E Aver(C, F). Conversely, aenrme that DiPr(C, F) Aver(V, *). Let D be any a set m C. For every p E 3, since (D,p) E Dist(Ct7), we obtai~ (D,p) E Aver(V,F). Hence, D below to VF. u

Fa most dasses C, we immediately candude the inclusion C E CF, since Dist(C,F) S Aves-(C,7)-

An advantage of Propodion 7.22 is that k ' s qnestion Dist(NP,P-comp) CAmfP,*) can be

simplyrephrasedasfoIIons "19NPindudedmP-?D Inanatte!mpttoanswerhis~nestion,we

mast make a careful study of the qnintegsentid c o m p 1 ~ dass Ph,. More genetdy, the qtrestion

Dist(4,Pcomp) CAtrer(AE, *) is transkted into the question EE C?A:-=~ in the worst-case setting.

h m Defhrition 7.2.1, we obtain quintessential complexity cIasses and XgF, where k E N Beedl the definition of the a- compledtg classes, Aver(dVP, F ) and Aver(IIz, F), k E N. These dasles

are cowentionally d & d to be the complements of the original defined-by-machine dasses Aver(NP,F)

and Aver(E:, F). Adapting Dehition 72.1, we are able to &definen the classes (~DNP)~ and IIEF7 Nev-

ertheless, do these classes conflict with the dehition of the comp1ement cIass ? More spec%cally, do they

equal the complements ofNPF and ELF ? The foIlowing lemma shows that W o n 7.2.1 does not eodict

wi th the complmed classes Aver(&, 7) in general.

L e 7 n ~ ~ 7-22? Let C and V be compfzity dasses and 3 be a set of d . .

1. If Aver(C, F) is defb.red, then a d F = (c0-C)~.

Proof. (1) For any set A E dFv we have 2 E CF. For every distribution p E 3, (x, p) E Aver(C,F).

4 our assumption, this is equident to the statement that ( q p ) € A ~ v e r ( d , F ) for every p E 3. Thus,

wehaoeA~(c0-C)~.

Conversely, assume that A E (WF. We have (A,p) E Aver(&, F) for all distriiutions p in 3, and

thus (z p) E Aver(C,F). Hence, a E CF. In other wordst A E d ~ . (2) By a similar, simple argument. a

In particularv we have the m e s IIiF = tmCEF and X S n II:F = (EE for dl k 2 0 and for

any set 3 of distriiutio~~~.

We make the fernark here that idthough we have seen in ComIlary 3.5.18 that Aver(P,P-comp) P Dist(NP, *), we do not know whether PWP P NP.

The foUowing indusion fb11ows immediately kom Coroky 3-47.

W e remark here that pHF is not dehed a9 the union of dl EEF 6X any k E N Hence, despite

Proposition 6.46, we m o t simply conclude that CPF = ELlF implies PHF = EEF. We next locate the newIy d&ed classes Pr, NPF, BPP3, and PSPACET in the worst-case world

3. BPP E BPPp E BPE.

I. PSPACE G PSPAC& c ESPACE.

Proof. We give only the proof of (I) since the rest of the cIaims foIlow by a -sirnilat argument. Since

Dist(P,F) E Aver(P,F), we have P E Pp. To show that PF C_ E, we Id A be any set in PF. Smce

(A,v-d) E Aver(P,T), there exist an increasing poIynomial p and a deterministic Thing machine M 1 which is ptime bounded on v-caverage such that M computes A. Recan that YWd(z) > 8(t,l+l)=-21.[.

It cIear1y holds that, for almost alI z,

for an appropriate consbnt c > 0. Therefore, A bdongs to DTRV~E(O(~~)), which is a subset of E, n

As a final note in this introductory subsection, we wish to demonstrate that the ciass %al P under FLAT" is not a Iarge class, where FLAT denotes the collection of all flat distriiutions. We note that

whether NP PmT is an open question, which is related to the incompleteness of distributional problems

with flat distriitrtion~~ We now rrcall that SUBEXP = n,,, DTIME(~(~~ ' ) ) .

Proof. Assume that A is in PmT. Let p be a distribution such that ji(z) 5 2-lrl*'"" for almost all

z, where e(n) = l/llogn]. This distriiution is flat because e is decreasing. By the choice of A, (A,p) E

Aver(P, *). There is a deterministic 'hing machine M computing A in time p on paverage, where p is a

function from E' to P. Choose constads c, k > 0 such that p(z) c - t ) for almost a11 z-

Let rn be any pasitive integer. Then, for any s&ciently Iarge z,

Thus, A E DTME(o(~R"".)). S m rn is arbitrary, A A E E X P htiows.

7.25 Red PoIynomiaETiie Hierarchy

The notion of computability enables us to translate all- complexity dasses into

wcmtase compIexiQ classes- Tn particular, we can natarally trarrslate the average polynomiaI-time &a-

arc& into its - eoanterpart, We caIl such a hierarchy the d poEynomia-- hiendqt. This

s a b s e c t i o n r a i l l s m d y i t s ~ p r o p e r t i e s

TheSwmaldehitbnoftherealpoIyn~timehierarchyisgivenbelow.

Deenition 73.7 (Beal Po1ynomirrl-Tie Hierarchy d e r 7) [97] Let F be any set of distribn-

tions. The d pol@umid-time hiamJIy rmda 3 consists of hiF, E!F, and IIiF for all naturai numbers

k. Let pHF be the coIIection of all sets A such that (A, p) is in Aver(PH, F) for all distributions p E F-

.4aaogoas to the real polynomial-time hierarchy, we can d&e the raPl plynomid-time o[kmatjon

hiermchg using the average po1yno~-time alternation hierarchy.

Deanition 7.2.9 @teal Polynominr-Time AUe!rnation HIerarcby under F) Let 3 be any set

of distriiutions. The d polywmid-time oltaMtion Idamdiy under 3 consists of AE&, and for dl k E N Let AP& be the cdeetion of all sets A such that (A, p) is in Aver(APH, 7) for dl

distnitrtiom p € 3.

Because of domination conditions imposed on orade 'IIPiag machines, the real polynomial-time hierarchy

forms a subhierarchy of the heas expond-time alternation hierarchy, {A:, q, II; I k > 0).

PmoE Recat that Dist (Ai,F) Aver(AAE, 7) and Dist (Xi, F) Aver(=:, F) for all k > 0. Hence,

A: AA:F and C: The indusions AAS E Air and ACEF CEF come from CoroIlay

6-53. In the following, we show the rest d the daim, namely, AEF A: and X& E q. It suffices to show that (4~-) E Aver@:, *) implies A Q qe This proceeds by induction on k. The

base case k = 1 foIloas b m Lemma 725(2). Let k 2 2 Let (A, b), 1 5 i 5 k, be a distn'butional problem,

where At = A and pl = v-. Assume that (4, h) i) Ave!r(NP, *)(Ai+x*fi+l) via an oracle machine Mi for

alliwith 1 sick. AssmnealsothatA&reeogaizedbyaT'urhgmacbineN whichrrmsinpolynomial

h e on Pk-a- gk(2) = TmelV(z), gi(z) = -&,.&+t$) & Q ( ~ ~ A+Iss) gi+l(z)- It not

Wdt by induction on i to show that is p o l p o d on pl-awage, sinQ gk is p l y n o d on pk-aiemg~

F i w e a m s t r u c t ~ ~ m a c h i n e s ~ , l ~ i ~ k,asintheproofthatAE~=X~. More

preciseIy, the m=uhine M . is d&ed as Mows on input z, M: sbdates in an Bdsteatid state Mi on the same

inpat except fcn orade queries; if Mi qnezies z, then MI guesses its orade answer m ( z ) and matinae the

~wilMireacheanaceeptingeonfigacation,k~dysimalatcsM&, onaIlzaithmrt(z) =O;

then in a rmiversaI state, Mi' simnlates Mi+l on a11 z with m ( z ) = 0; M{ enters an accepting configuration

exactIy when ML, h e s a rejecting configuraton. Otherwise, M{ enters a rejecting codgmatioa It is

obvious that each Mi' is an alternating Thing machine with (k - i + 1)-alternation. Moreover, assuming

x E L(Mi), the length of the minimaI subtree of the computation tree of M,! on x which contains ody

g p e s " - c o ~ t i o n s is bounded by c -gr(x), where c is a positive constant not depending on the choice of z.

Seoond we defiue a new machine M as follows: on mput z, M simulates Mi on z in time 2elrl*; on each

computation path, if Mi does not reach any halting configuration, then M enters a rejecting dgurat ion.

Smce is polynomial on v-d-awmge, z E A if and only if z E L(M). Hence, we have A E q. The case for AEF is shown analogonsly. 0

In worsecase compIexity theory, PH is the union of all sets in CE for any k > 0. NevertheIess, we have

no proof that pHF = Uks CEF. On the other hand, can we show that PHF differs from Uk, C ~ F ?

Smce it is still possible that PF = PHF, the separation between PHF and Uks C i s seems dif5cult to

prove. This situation is s;mitat to the question of ATIME(~O(')) versus u ~ , ~ A T I M E ~ ( ~ , ~ ~ ( ~ ) ) . Because

Uk>o ATIMEZ(k, nO(I)) coincides with P H while ATIME(~O(')) eqnals PSPACE, we do not know whether

P H = PSPACE. Therefore, as mentioned before, we cannot conclude that if Cir = C;+p, then PHF =

q F . We have seen that A: = CE impIies Aver(AAE,F) = Aver(AE;,F). Hence:

Last, we shall demonstrate some basic closnte properties7 under set operators, of cIasses m the real

polynomiaI-time hierarchy under P-comp.

h E Let k > 0. Here we shall show the closme property of XiF under @. Let us assume that A and

B are m XiF: that is, ( 4 p ) E Aver(Ci, P-comp) and (B,p) E Aver(I=E,P-comp) for all distniutions C( in

P-comp. Take an a r b i i distribution p E P-comp. We assume that b(A) = 0 for simplicity. Let g and

a be d&ed as

where b E {O,l). We next de6ne p' as br(A) = 0, and for each nonempty z,

Xz = 6z for some b E {O,l) and z E Cf, i€r=bE{O,l).

he &dm p' becomes a distribution and satisk jI'(0r) = @'(CP) = 4- It is also easy to see that

<~C(becauseb(z) 1 2 - P ( x ) . Hence, (A@B,P) 9m(A@B,Cc3- P-

D e h e ji~(x) = 2 ji'(0z) and ji&) = 2 - jit (lz) for all z. Notice that p' = p,.t @ pg . By the

definition of pt, both and pg are PampufabIe. Thus, we conclude that (A,pA) and (B,pB) are in

Aver(C, P-comp). Lemma 6.4.7 yields the conclusion that (AW, p ~ e p g j E A W C , Pamp), which means

(A $ B, p') E Aver(C, P-amp). Since Aver(C, Pcomp) is dosed under I%-reductions, then (A 63 B, p) is in

Aver(C, Pomp).

Since p is a r b i i , we get A @ B E C-,.

7.2.3 Nearly-Ci and Nearly-A: Sets

In Section 4.7, we introduced the notions of "nearly-RPn and "nearly-BPP" sets. In a similar fashion, we

can extend this notion and introduce the new notions of "nearly-Ef;" and "nearIy-Ap sets. Out goaI here

is to prove that every set m ZiF (AET, resp.) is nearly-Ci (neady-A:, resp.).

Definition 7.2.13 (Nearly-Ci and Neatly-A: Sets) A set A is if, for every p o l p n d

p, there exist a set S and a poIynomial-time alternating Thing machine M whose alternation is at most k

starting wi th an sdstential state such that (i) z E A - S implies M(x) = 1; (ii) x E x- S impiies M(z) = 0;

and (iii) R,[z E S] < for almost all n. smilarly, a notion of *hemlpAin is de[ined by using an

alteroating T m k g machine with a semideterministi c process.

PrqpositiOn7.21~ t e t 3 k a s e t o f d i s k r I b u t i m s t r c l , t l u r t v ~ ~ 3 . LetCbeuneofthefdlotaing

classes, A;, Z:, k E N, BPP, andRP. Then, errerg set in CF is nanfy4.

Roo£ F i we s h d show the case C = E;. The other case C = AfE f o U m s;mnatIy. Let A be

an a r b i i set m Xif. Note that (A,vsrand) E Aver@:,*)- By Proposition 6.4.12, we have two sets

Co E EiandC1 E I I i s u c h h t Co C A , CI C & and I ( G u ~ ( [ 2 (1 -l/p(r~))-2~Coralln E N LaS=F-(C0uC1). ~ e t h e n h a v e p < ~ f ~ a l m ~ s t d l n , a n d a I s o w e h a v e C & = sand

Cl = 2 - S. Thedore, A is m nearly-Ci.

Next we shall prove the proposition fix C = BPP. Assume that 4 f BPPr. Smce v- E 3, we

have (A, V-) E Aver(BPP, T). For OOIlveOience, mite w for v-d. By the d-on of Aver(BPP, F),

there is a bormded-etror probabilistic Thr& machine M computing A m time p on waverage, where p is

an incfeasing p o l y n d In particular, for the randominput domain rM assodated to M, iz., (((2,s) / TiieM(z;s) > p(lzl r))) < l/r for any real number r > 0. By Lemma 3321, it foIlows that B((z I Tirme&(z) > ~(1x1 -r))) < 1/r. We can assume by the Amplification Lemma that &[M(Z) = A(z)l? 5 /6

for all 2.

Take any polynomial q, and let q'(n) = p(n - 40(n + I)* - q(n)). We then d&ue

FlrSt we shall show that the density of S is not so kge. Fix n E N.

Next we define a randomized M g machine N as foIIows: on input z, simulate M on z in time q'(1zi);

if a computation does not terminate, then simpIy reject the input. Note that the error probabi i is less

than 1/6. We shan show that N meetly computes A on most inputs. For s not in S,

Hence, A is nearly-BPP. The case for C = RP is similar. o

Roof. Suppose that there is a strong one-way fnnction. Asme also that NP E BPPF for some F with v-d E 3. By Proposition 72-14, every NP set is nearly-BPP. Proposition 2.6.4 shows that there is

no strong one-way hction, This is a contradiction; hence, NP BPPF. a

7.2.4 Collapsing Classes

Let us return to Levin's or ig id question d whether NP E P-=,. As diseossed in the previous section,

we can now raise the more general question of whether Xi E A$ holds for some F. CIesrIy if P = NP, then NP is indtxded in Pr for an 3. BemDarid d d [9] first gave a partia answer to this question by

showing that Dist(NP, Psomp) A W , *) if E # NE. In other words, NP C PmP implies E = NE Note that E = NE if and d y if TWY n NP P [14]. Hence, Ben-David et d. actuaIIy showed that m y n P M p CP.

We shall generalize this result and show that the M y part of any red average m m p I e dass dapses

to its worstase colmterpafir Fnst we state a technical Iemma R e d the standard distn'bution uuY. In

the rest of this section, we use this distriiution,

Roo& Let A be in TALLY n CF. W e note that (A, uuy) E Am(C, F) since urny E 3. By the s p a ~ e

interpol;itim property for Aver(C, F), there exists a set B in C such that An (0)' E B E A. Since A {O)', wehave A = B. Thus, A belongs to C. R

R d that most average-case complexity classes dimmed in this thesis enjoy the sparse interpolation

property. In partidar, we ha=

S e n d corouaries follow h m Proposition 72-17.

Root The daim f01lows h m the fact that TALLY n A: 'j A: but TWY n Ag E A:. The cases for

ZEF and IIrF are simil=u.

It is anlikely that A! = q; thus we may conjecture that XI AiF for dl 3 which contains u w - In other words, the sets in XE seem hard to compute even on average

When we consider the daPs PMp, the daim of hposition 7217 can be strengthenred in the foIIoning

manner- Recall that a set S is P-printable if and only if there exisfs a poIynomial-time comprrtab1e function

which, on mput On, Iists an strings in S of length r~

R o o E Let A be an arb- set m P- Take any P - p w e set S. Under om a~fltmption, we

map assume that there exists an integet k > 0 such that IIS n PI[ = (n + I ) ~ for dl n E N, since if the

demitp- of S n F does not reach (n + Qk then we cau dete . - ticany add an dement, which is not m S,

to S repeatedly d its density is exactly (n + 1) k.

Let us dehe b(x) a (Iz[ + I ) - ~ - ~ if z E S; else 0. Notice that p is P-computaMe Take a machine

M which computes A in pcdynod time on paverage. There are integers m,c > 0 such that b({z I Ti ieM(x) > c(lz1- r))) < l/r for all r > 0. For any fked z E A n S, we have Time&) 5 ~(Jzl / j i (x))~ I dlzjm(]zl + 1)'D(w2) - Hence, A n S E P.

T h e m halfofthechimfotlmfrom the fact that allsparsesetsmp areP-printableifP=NP- 0:

We have seen that the tally part of quintesSentiany computable dass CF is easily computed, and thus

it collapses to its worstcase counterpart C If we take the set of recllrsive distrtitrtions REGcomp, then

CREC'COmP mUapSeS to C. TO prove this? we show that if Aver(C, BECcomp) has the sparse interpoIation

proper& then Cmcomp C. In the proof of the foIlowing lemma, we again use rWee? recursive, proper

hard cores (see Definition 25.1 1).

Proof. Suppose that A M , REGcomp) has the sparse interpoIation p r o m - The proof is by contra-

diction. Now assume that them exists a set A in C--=, - C. By Lemma 2-5.12,'there Bdsts an infinite,

recarsive, proper hard core H for A with respect to C. We note that if C = P, then H is in the class E (see,

eg., [4]). Thus, for any set B E C, if B E A, then B n H is finite. Now let S be a renndve, W t e , sparse

subset of H. Let q(n) = IIS n Fa. Consider the distriiution psVq sacit that

3- BPPREC-~ = BPP.

has the sparse interpolation property- The daim for C E {BPP, PSPACE) foHm h m Reposition 6.43,

and the claim for C E {A:, El I k > 0) foIIows h m Proposition 6.4.11. 0

Theorem 72.23 indicates that the de$nition of the average polynomial-time hierarchy in Section 6.4 is a

reasonable generabation of the worst-case polynomial-time hierarchy.

Note that? in the proof of Lamna 72.22, the complexity of the distribution psBq depends only on the

comp1exity of the complexity core Since aII sets not in P have comp1exi~ cores in E, we get the foUowing

coronaly.

Complex distributions Eke recursive distributions make quintessential complexity classes lose their average

case nature. This supports our primary interests in feasible distniutions in Chapter 4. In later sections, we

shall f m on qnintessential complexity cIasses under sets of those feasliIe distriiutions.

7.3 Fundamental Separations

We have seen the collapse of the real polynomial-time hierarchy under the set of recursive distributions. This

section shows the sepantiom between the real po1gnornial-time hierarchy under Pcompptrtable distn'butions

and the polynomial-time hierarchy- The technique cultivated in this section is frrndamtal and will be used

again in lateT sections.

7.3.1 Construction of Hard Instances

Three years after the notion PmmP was pmsentd, Schder [92] succeeded in showing that Pkm, # P by constructing a complex set which Iies in the diikeace P h m p - P. A crucial idea in his proof is to h d a

string of earh length which oeaas with low probability by pruning other strings which occur wi th relatively

high probability- The construction needs aa effe&ive enmneration of P-computab1e semi-distn'btxtiom. Later

Schaler and Yamakami 1981 extended this d t to create sets which are hard to compute even by O(P>time

bounded Zitring machines for a eanstant c > 0. CertaWy we cannot extend this r d t to p(n)-time

bounded machines because all in PMP are already in E. Hence, this resnlt seems nearly optimaL

The following lemma w i l l be d m kter sabsectio~ The proof @wen here uses Lemma 42-15, due

to Sfhaler [94, which uses resource-bounded KoImog~f~~ c o m p I e to avoid any enumeration of semi-

distributions for the constrnction of a hard set.

For a set A, define Prefiz(A) = {On1o I 3 w ~ w l = n A w E A]).

Proof. Without loss of generalitp, we may assume that s(n) 5 Iogn for all n. Since k(n) E n(n), we can

take an integer q > 0 such that k(n) 2 2 h a l l n E N Let l(1) = K(1) = 1. For n 2 2, Iet l(n) = 110 lognj

and let K(n) = max{l, Lk(n) /10 - log nl). Now consider an integer nl > 0 such that n 1 40cl - log n for all

n > nr. Obviously, K(n) - I(n) 5 k(n) for all n > 0. Note that, for all n 2 nri Y(n) + l(n) 1 $k(n) Siflce

k(n) 2 rll) log n and, in consequence,

It is also clear that kJ(Izl), k.k([zl), and kK( l z [ ) are unbounded and compntabIe in poIynomia t h e

since k(n), s(n), and An. LlOIognl are timeconstrnctibk

To simplify the following aaalysis, we always disregard the computation time for the dues k(n), t(n),

and k'(n) because, as we have seen, Xzh(ld), Xz.s(Izl), and Xzi(Iz1) are aII P-computable and do not dect

%verageD nmning time

As for the desired het ion h, we define h(ln) to be a string $ - - - z&,) gL(n)-k'(")'(n), where ,Zt =

min(w E I ru ~'T'[n~!i'(~)q) and 4 = min{m E e(n) 1 w $? KT[n,Z'(~"'~r~---z~-J) for all i

with I < i 5 K(n). Obviously, we have Ih(ln)[ = k(n) for all n > 0. Now define the desired set A

as A = {no I 3n[z = f(ln) A ut E !P-lzl A n > 01). Note that A n E" = h(ln)p-lh('")li and thus

llAn !P(l= P-Mn) for at n > 0. We next show that A n B E MiPtomP for aC1 sets B in A:. Now &c a

set B in A;. Consider the following algorithm M:

begin dgorithm M for A n B input z (let n = 121) compute k(n), s(n), L(n) , and K(n) (m polynomial time)

I e ty=X

for i = 1 to K(n) d o

compute $ = min{w E Et(n) [ w 4[n, 2'(ga Ir])

let y = y g

i f y g z then* end.-fiIu

It is to see that the algorithm M computes A n B. For bter convenience, we calI a sbing which goes

into the line (*) a candidate.

Now let p be & P-eomputable distribution. By Lemma 4.215, there exists a mustant Q,M > 0 such

that, for all i with co < i 5 K(n), b(zf - -z~C"-~'(*)) < 2-('-~)'(*). Let d be a iarge mteger so that, in

the following estimation of the time mmplexity of the algorithm, the inequalities (u) and (* * *) below hold

for any integer n > 0. Moreover, we assume that the computation time of the set B on input of iength n is

at most F'-~*' for each n E N Let nz be a constant large enough so that nz > max{%, nl), n 1 W , and

n 2 k ( 1 0 ~ + 1) log n + 1 for all n 2 nr . In what follows, we assume that n is any integer Iarger than nl.

Let Zi be the set of strings of length n which are not rejected through the first i iterations of the for-Imp;

namely, Zi = Z: - - -z~P-~'~(~)- By dehition, Zo = C" and A ZuCn). RecaIL that ji(Zi) < 2-('-")'(").

We then partition 2' mto K(n) +2 subsets, S = {Zc-r - Zi I 1 5 i 5 K(n)) u (Zu(,) -A) u {A). Note that

K(n) + 2 5 n since n 2 n2. By Lemma 33.15, it a c e s to show that, fa some constant c > 0 independeat

of the choice of n, Thne~(z) I nc + (l/n2ji(D))e for a11 t E D, where D is an arbitrary set in S. Note that, to compute zr, m the worst case we have to check alt strings w m c'(") whether w 4

KT[~, 2'(')" I y]. Hence, we need at most 2'(") . 2'mL(*) < - 2kH1)l(") steps because log j 2 s(j). Smce

i l(n) + 1 5 y("), the running time of M on input s which is in - Zi reqtiires

since l o p 2 Iogd + 1, Iogn 3 logi, and I(n) 3 2logn.

Let d be the minimaI integer satisfying the following conditioa- i 2 logi + co + 3 for aII integers i > d

Note that d does not depend on the value of n. Let c = max&d, [5(Iogd + 211)- We examine betow the

nmning time of M for several cases of input string z-

(i) For the computation on all rejected shkgs z in C - Zd, fir i < 4

(ii) For the computation on all @ected shings z in 2'-1 - Zi with d < i < V(n), as we have seen,

TiieM(z) 5 2h'+2)'(n). Note that, since i 2 logi + Q + 3, (bgi + 2)L(n) < (i - co)l(n) - 2 logn Then we

have

(iii) For the computation on all tejected strings x m Zu(n) - A, note that zr zE(,) x. We dso note

that k(n) - K(n)t(n) - 1 5 wk(n)+L. Hence, the time spent for z is

(in) For the computation on aiI strings x in A, we remark that the computation of B needs, by our as-

sumption, at most @-n+c' time Hence, Tie&) < F'"+~. Note that b(A) 5 fi(Zr(,)) < 2-(Y(")-~)t(n).

Reean that K(n)t(n) 2 ik(n) and k(n) 2 $. We Re remark that dn+d 5 e((K(n)t(n) -a) t(n) -2logn).

This is seen as follows: since c 3 4c'~ and n 2 2cl (lOm + 1) log n + I,

n + l n 5 --- + 1 0 ~ - log n

4cr 2c1

Therefore, we have

By Lemma 3.3.15, we condude that krcTi~neM(z) is polynomial on paverage. Smce p is arb'rtrary, AAnB

belongs to AAEkmp.

The same argument can be d e d out for the daim (4).

Next we show that cIaim (5) also holds. To prove (S), we consider the foUowing algorithm Mr.

begin algorithm M' for Prefiz(A)

input z

b d u such that x = 0~-1~l lv

if no such v exists then reject

let y = X

for i = 1 to K(n) do

compute z$' = min{w E z ~ ( ~ ) 1 w p K T [ ~ , ~(g*(y])

let y=&

ifugythenaccept

if y o then reject

Thaorwn 7.3.2 Let k > 1 and let c & any positive real nuder.

1. SPARSE n M;P-, g ATIME~(~, O(yn)).

2. SPARSE n A.E~p,,p -(k, 0 (2~") ) .

Proof. Let c be 6xcL

(1) Let (Mi)iEN be an & c t k enumeration of a11 semideterministic dternating ' h h g machines

which work with at most k-dtemations in time O ( P ) , with W y many tepetitions. Now we de-

fine the diagonaliud set D = (z I z 4 L(MLlwI,II)). It is not hard to see that the set D beIongs to

ATIME~(~ , 0(2(*l)")), but not to ATIMEA(k, O(yn)). By Lemma 7.3.1, we have a set A of density 1

such that A n D E We show that A n D 4 ATIME~(~,O(P)). Assrnne that there exists the

ith machine M, which computes An D. Let z be any string z of Iength 2'. Then,

This is a contradiction. Thus, A fl D # ATIME~(~, 0(2")).

(2) Consider a set A and a fnnction h as de6ned in Lemma 7.3-1 such that A n C = (h(ln)) for a11

n > 0. Recall that A beIong to D T I M E ( O ( ~ ~ ")). By an Wemation" version of a time hierarchy

theorem [99], there is a taIly set B in the d i f f e ~ e ~ l e m ( k , O(F -log n)) - - e ( k , o(P)). Define

C = {z I z E A h 0l21 E B). Clearly C E XiPTOmp by Lemma 7.3.1. Now we prove that C is not in

ATIME~(~,o(~c")). Amme otherwise Consider the following dgorithm that works on (0)':

begin alternating 'Ihring machine

input on compute h(ln) if h(ln) E C then accept else reject

end.

Smce A n E" = {h(ln)) for all n, the abcm aIgarithm computes A. The rrmning time of the algorithm

on input On is

for any & d d y Iarge integer n, where C is an appropriate positirre axmant. Theref' we have B E ~ ' ( k , 0(2~")). This is a contradiction- a

Now recall that At, = ~-(k,*(")) and = ATIME'(~,~~(")). Since Mi-, A; and

ACE-, E Ce,, the results of Theorem 732 are nearly optimal; thus the classes A;-,,,, and Cg-, share hard sets with A: and Fk. As a particular case, the class P-, has hard sparse sets in E. In light

of resourcebounded measure theory, however, the cIass DTIME(0(2Cn)) is known to be small within E, and this suggests that the class P-, may not be a large dass within E. In the later section, we shall

show that P-, is actually smolL

Clearly P E P-, P k m p since P-comp C P-samp. Can we show that P-, # P ? In contrast

to Theorem 7.36, if P-, # P, we can solve some open questions m worst-case complexity theory.

P t o p o ~ 7.84 If Pp-, # P, then either #P or NP BPP holds.

Proof. We shall prove the contrapositive. Asnrme that F P ~ E #P and NP BPP. Sice P-samp c #P-comp, we have P#-,, 5 PMp. Under the assumption NP C BPP, by Corollary 4.7.7, P-, = P#-,,. Oui assumption FPE E #P leads to the condadon that E a m p C #P-comp since

Hence, we have P#-, E P w - Recan from C0roh-y 72.24 that Pkmp = P. This yields the

desired condusion that P-, = P.

Here we show another application of Lemma 7.3.1.

Deenition 7.3.5 (Sparsely Close Sets) For a c o m p l e cIass C, a set S is d e d (qmsdy] C-Jose

if there exists a set B E C such that AAB is sparse. For the sake of convenience, we aIso use the notation

C-cluse to denote the collection of alI Cclose sets.

Note that any recursk set whase polynomial cornp1- core is sparse betongs to P4se

Pmpodkm 7.3.6 Let k > 0. For any c > 0, AA:P-mp A T I M E ~ ( ~ , O ( ~ C ~ ) ) & ~ ~ . In mculat,

Pp-eDmP P - c l e

h E The proof is by diagonalization. Let C = {z I z $! L(Mlh where {Mi)im is an dedive

euumeration of semi-ciderministic k-dkmation, O(P)-time bounded alternating Turing machines wi th

infinitey many repetitions. We have C E ATD&(O(~(*~)")). Dehe A as in the Hard Instance Lemma

bychoosingn-wnl ask(n). It f o I I o t ~ s h ~ ? . 3 J t h a t A ~ C ~ A ~ p - e 0 ~ ~ -

We next show that A is not spame Note that ll(An C)M(Mibnl) n FII = IIAn FII- The density of

A is, for each n,

Hence, (A n C ) M ( a ) , i E N, is not sparse since infiniteiy many machines coincide with Mi. 0

Recall that NT is the condon of afl near-testable sets. It is k n m that P NT E n PSPACE [31]. We show that P h m p contains a non neareartestable set.

ProposiEiolr 7.3.7 [98] PP,, P NT.

Root D e f i n e A a s i n T 6 e o m 7 3 2 ( 1 ) v i t h d ~ ~ b y iIAnC"II =l foreachn>O,sothatA

separates Pmmp h m DTIME(O(P)). We show that A is not near-testable. Amme otherwise Let 6arndmy(A) = { x I z # A, d y one of x and z+ is in A }, where z+ is the

successor of x. Notice that bamdmy(A) = ~nmdory(;i). We daim that A is near-testable if and ody if

b t m h y ( A ) is in P. Define f by f(z) r xA(z) ~ ( x - ) (mod 2) for alI x. By the definition fo f , f (x) = 1

if and only if exaaly one of x and 2- belongs to A. This yields the daim.

Thus7 we have h n & y ( A ) E P. Note that the set A n C contains at most one string, i.e., A A SPARSE. Consequently, A is in P. The eonclasion that A E P contradicts the fact that A fZ DTIME(O(Zn)).

However, it is open whether PMmP E $P or not.

7.3.3 Separation &om Advice Hierarchy

Another immediate consequeuce of the Hard Instance Lemma bdow is to show a relationship between

and the dss AE/cn d&ed by linear advice frmctions (see, eg., [4,41]). Fm we introduce an

advise hierarchy m the following general fashion.

Debition7.3.8 (CwithAdvicef) ForCacoanpl~daPsaadfafim~onhmNtof',a~Sis

inC/f (Cnithadvicef) ifthereisaset B E C S U C ~ that Sn!P ={ZE F I (x,f(n)) E B) fordintegers

n > 0. For a class 7 of frmaons h m N to C, let C /F = UIEFC/f-

For example, the c b s P/poly is known as the dass of sets computable by non-dorm fhdies of

polynomial-size cireaits. For the g e n d properties of the advice hierarchy {Af;/poly, EL/poly, IIE/poiy I k E N), see Yap [fill. Note that whether X: A//pIy is an open qnestio~

We want to show that A:- is not included in AE/cn when c is hd

Root Again we use a diagodization argument. Let A be the set m ATIMEA(k,O(n~'")) whose

existence is gwanted by Lanma 7.31, such that [Anll = 2Lbnzl and A n B E A:- for every set

B E A:. NOW let {MiliEn. be an e k t b e enumeration of all p d p d t l m e semi-determiniaic k-*on

bounded alternating 'Ihdngm=uhines7 eachofwhich, Mi, nnrr in nitisteps. We J6all dehe aset which dbg-

CaAPTER7. QUINTESSENTIAL COMPlXMBEDY 253

Let B be the set accepted by the above 'Iining madhe. Since B 2 A, it d c e s to show that B E A&

For each input z in A, the machine takes 0(2(*')") steps since it needs O(2" - n'q* ") steps to enumerate

all elements in A n P, at most n2 iterations of the for-loop and O(n'wn - 2C") steps to compute each

ACC(Di, q) and REJ(Di, zi) fm some constant c > 0. Therefore, B Bongs to At,.

We show that B f A: / cn Let us consider the set {Di 1 < i < 2Lkn2J}. We remark that the definition

of sets Di, 1 5 i 5 2LfOf5n2J, does not depend on the choice of sthgs z in A n IF. Consider the IMX~WI

k such that Rk # 0, i.e, the algorithm goes into the line (*). Note that k exists and k < cn- By om

definition, either ACC(Dk,zk) or REJ(Dk,zk) is empty- By (*), ACC(Dk,xk) # 0 d y when zk fZ B holds. Hence, there is no advice string z E ECn such that zi E B if and ody if (G, z) E L(MlIognJ). U

Despite of the above theorem, it is still open whether A: /ply. Schder [93] p r d

negative evidence by demonstrating that if P-, E P /poly, then EXP = qP. Here we wish to prove

that A:-, A: / p l y if and ody if A? G A: fpoIy.

We shall show that every AF-set is reduaile to via po1ynomiaEsize circnits. We &st describe

these reductions.

Dehi t ion 7.3.lO (P /3-m-redtlctions) Let h be a fanetion from N to Co. A set A is P /h-m-reddte

to aset B7dendedby~<Ih ~~if thereexistsafnnct ion~~~~suchthat ~={z1g((x,h(lzl))) E B).

For a set 3, A is Pf3-m-redtrdble to B, deaoted by A <IF B, if A B for some h E 3.

Note that A <g B and B </O(nl C imply A &/po'y C, where pdy is, as More, the set of pbounded

functions h m N to C*.

kmma 7.5.11 For each k E N, the classes AE/poly and C:/poIy me closed dourmrrard under P/poly-rn-

redudr-m.

R w E We shaU prove the lemma only for A:/poly. Asmme that A B and B E A//poly. Since A I ? ~ ~ ~ B, there exists a function g in FP and a pbounded funaim h from N to C* such that

A = (z I g((z, h(lz1)) E B). Also by oar assumption that B E A:/poly, there are a set D E A: and a

pbonnded function f from N to C' satisfying that B = {z I (z, f(lz!)) E D).

Combining those two set equations, it foIIows that

To show that A is in A;/poly, we define the set E as E = {(z, y) I @((%, (y)~)), (y)~) E D), where (y)o and

(Y)I are decohgs of Y stisEPing Y = ((do, (Y)x). Let us set q(n) = ( W , f (4) for =tual numbers n-

Clearly q is pbotmded becanse both h and f are so. The d a o n s of E and q yields the desired condusion

that A = (z I (z,q(lzl)) E D), which implies that -4 E Ailpoly. o

The following lemma generalizes the s p d case (k = I) proven m [93].

k n m a 7.5.12 Let k > 0. Euery set in A? is P/poly-m-red& to some set in A:p,,,- A sirnilat

daim holds for C F and x;~,,~.

ProoE Let S be a set in A?. Note that every Ac;" set is pm-redna%le to some set in A; by Lemma

25-10. Hence, there exists a set St m A; such that S <k SS. Let us define the set L to be the conection of

all strings z such that z = zy for some y E St and z with Izl= [yl+ b, where b r 1x1 (mod 2). Crearly L is in A:.

Take a set A and a function h dehed in the Hard krstance Lemma with the condition Ih(ln)l = [n/21.

She L E A:, the intersection L n A Iies m A:-,,. Now let T = L i7 -4 It foIlows that? for all y of length /W'Y T* R Y ~ ~ t Z a n d o d y Z h ( l n ) y ~ ~ . H ~ s & / ~ ~ ) T . ~ i c e ~ g ~ t , w e h w ~ ~ g o

Using the above lemma, we can show the intractability ofthe classes in the red poIynomiaI-time hierarchy

under P-comp.

Theinwn 7-3-15 Let k 2 I. LetC E (A:,C; f k E N)-

7.4 Immunity and Bi-Tmmnnity

Immttxte sets are another typical example of hard sets. In recursion theory, a set is d e d "imrmme* if it

is inhite but contains no infinite recursive enmnerabIe set. This notion has been adapted to complexity

theory and used in a variety of situations.

Bibimmune sets are d&ed to be immune sets whose compIements are also immune. These notions are

fundaments and have already appeared elsewhere in this thesis. In this section, we shall pay more attention

to the existence of such sets in P h m p .

7.4.1 Immune Sets and Complexity Coma

F i of a, let us r e d the formal dehition of immunity. For a complexity dass C, a set S is C-innmme if

S i s ~ a n d S h a s n o ~ t e s u b s e t s i n C .

A general notion of wmplQity cores has been introduced in Section 3.5, but in this subseaion, we shall

focus only on complexity cores with respect to P, the d e d polyunnial umrplezity cores. For a recursive

set S, a set C is d e d a polynmnial complQity am for S it, for any deterministic Turing machine M computing S and any polynomial p, the set {z E C 1 T i e ~ ( z ) 5 p(jzj)) is finite.

This section wdI show that there exist P-fmmuae sets in PMP, but P-,, has no P-bi-immune sets.

Under the assmaption that P = NP, ail infinite polgnomial complexity cores for sets in P-, are shown

to be hard to compute

We bt show the odstence of a P-immune set m P-, of arbitmy density. The proof bdow uses an

elegant technique deveIoped by KO and Moore [56].

ProoE Let {Pi)m be an dective enumeration of all sets in P, where Pi is deterministidy computed

inni+istepsl

Let c = (1 + 6)/2 and take the set A d&ed in the Hard Instance Lemma such that jlAnll 2 2C". Note

that A E DTIME(O(nlw. n)). Now consider the faowing aIgorithm:

accept

end- for

reject

Let B denote the set accepted by this algorithm and let C = A - B.

Finst, weseetbatCisofd~at least2% ~ m r m e z , ~ E ~ a n d z f ~ . N o t e t h a t z ~ e a n d p ~ Pi

are witnessed through the first for-loop, then i # j by the algorithm, Hence, we have llB n PI1 log' n-

S i IIAnPll Z P , [CnF l l = IIAnIl -IIPII 2 2 " - I o g ' n z P .

We next show that C has no inhi& P-subset, Assume that an idbite set in P is a m b of C. Let io

be any integer that guarantees that Ph is such a set. Consider the mhbal x E f$, fl& such that 1x1 2 io.

By theminhalityOfx,wehavez #Ph f o r d z wi thh 5 IzI a n d t <z. T h u s , x ~ B,acontradiction.

Therefore, C is P-immme-

Now we daim that B is in PkoP since this implies C E PkmP. Let TimeB(z) be the executing

time of this algorithm on input I. It stdEces to show that Tie&) ?c'" for dI x m A, where c' is

an absolute mve constant. Notice that TuneA(z) 5 ~ 1 ~ 1 ~ ' ~ s ' 1.1 for some eonstant c > 0. Thus, the

algorithm takes time - [ C I L ~ * ~ ~ ' 1'1 + (121' + i)] m the second for-loop, and this tenn is bounded by

0(n2 1 ~ * - Za) 0(22n) - The tota exeation h e , TuneB (z) , :), log' n iterations of the first for-hop,

each of which takes O ( p ) steps, and therefore, 5 2 ~ ' " for some eonstant d > 0. a

We note that any P-immune set is a polynomial complexity core for its& Since the P-immune set

constntcted in Theorem 7.4.1 is non-sparse, we imm-e.1~ get the following coroIIary-

Let m recaII the eomplexitp dass APT introduced by Meyer and Patemon [n]. A set S is in APT if and only if the set {z I Tie&) 2 p(lzl)} is sparse for some polynomial p and some deteTminiStic %ring

machine M which computes A.

Proofl This result foUm horn Comkry 7.42 and the fact that a recursive set S is m APT if and only

if any poIpmial complexity core for S is sparse [7 ' . a

W e have aIready seen the existence of pol- t-me sets m Section 42. The particular example

shm there was; based on KoImogorov complexity sets. Here we shaIl present another example of poIgn0-

d y t - r a r e s e t s b a s e d o n c o m p ~ m r e r i n P ~ .

CHAPTER 7. Q-COlldPmILITY 257

Proof. Let S be a comph&y core for A with respect to D T I M E ( ~ ~ ( ' ( ~ ) + ~ ~ ) ) n PP-co,,. To atrive at

a contradiction, we assume that S is not potynomiany t-rare. Fkom this assumption, it follows that there

exists a Pamputable distriitrtion p s a w the condition that the set B = (z E S - {A) I B(z) > ZC("))

isinhi&.

S i A E Pm,, there exists a d-c ' k ing madhe M which computes A in polynomial time

on paperage. Let p be a polynomial such that h .Ti ie~(z) is p on paverage. For any strfng x in B (say,

n = 14>? Time&) 5 p(Y(") - n) 5 (2'(n)+'wn)k

for some constant k > 0 independent oiz. Now we set B' = (z I T i e ~ ( z ) 5 ($(")+'w")~). It is clear

that 3' belongs to D T I M E ( ~ ~ ( ~ [ " ) + ~ ~ ) ) . Notice that B' n S is This obviously contradicts our

assumption that S is a complexity core for R Therefore? S is polynomiaIIy hare. 0

7.4.2 Bi-Immune Sets and Resourcd3ounded Measure

We shan turn om interests to bi-Immune sets. A set S is Caned C-bi-immtme if S and its complement 5 are

both C-immune. The dass of P-bi-immtme sets has a dose connection to fesource-bounded measme theory.

As seen in Proposition 2-7-11, any class which has no P-bi-immtme sets has pmeasnre 0; in other words, it

i s d

It is known that the set E has %tm@f' P-biiimmrme sets 151, and thus E has P-bi-immune sets. It

is important to remember here that a reansioe set S is P-bi-immune if and only if 2' is a polynomial

compIsdty core for S. Hence, E contains a set for which Em is a c o m p l e core- However? we can see m

the following proposition that there are no P-bi-imnmne sets in PMP- This contmt dearly shows the

di&rence between E and P-. W e r and Yamakami 1981 6rst showed that PMP has no P-bi-imrmme sets. Later Schder I941

extended their d as follows.

ProoE. Assume that B is DTIME(O(2en))-bi-' i and 3 is ptt-redtxa'bIe to a set A in Pkmp.

Whout loss of generaIitp7 we assume that c is a positive'hegm. Let Y be a ~~~~~time o r d e Tbkg

machine which reduces B to A with nonadaptive qneries. Let p be a poIynomia such that n ~ ~ l e ~ ( Z ) (= p(I4) foranz- S i A E P ~ , t h e r e i s a ~ c ~ m a c b i n e N w h i c h c o m p u t s Aintimepolpnomia

on p m m g e for every P-compntabIe distnbntion p This shall Iead to a contradiction-

W e d e h e a s e t D a s h h m Foreachn E& kt&= {yI lyl ?n/!kl\y €Q(M,W)),andthen

C&QPTER 7. Q- COMPUTABILITY 258

let D = (0)' U (Un, D,). Fi we shaw that D is P-printaMe. This is seen as follows. Consider the

aIgorithm: on input On, recursively take a natural number k such that k 5 2c - n, and list dl strings in {y E F' I y E Q(M,P)). S i M makes nonadaptive queries in polynomial time, this algorithm writes

down all query of length n in polynomial time. Now let us consider the set A n D. By Lemma 7221, A n D must be in P. The FoIIowing algorithm

00mp~tes B n (0)':

begin algorithm for B n{O)' input on list an queries made by M on mput On

for all query string y

i f l y l 1 h thensetans(y):=[y~ AnD]

(*) else SimuIate N on input y and let an&) be its output

end-all

simdate M on On with oracle {y I cms(y) = 1)

output M(0")

end.

In line (*), the number of steps we need is at most 2elol 5 2~e = 2"j2. Hence, the total number of steps of this aIgorithm is, for some absolute constant d > 0,

fix every sutEaently large integer n. Therefore, the set B 17 (0)' belongs to DTIME(O(F)). A similar

argument shows that B n (0)' is m DTIME(0(2n)).

Notice that at least one of the sets B n(0)' and Fin (0)' is infmite This contradict5 our assumption

that B is DTIME(0(2M))-bLimmnne a

For each constant c > 0, the class of Dm(O(2C"L))-bi-immune sets is known to have pmeasure 1 [HI. In other words, the class of non-DTIME(O(2C"))-b'bimmme sets has pmeasure 0. Eence, by Propositions

7.45 and 2-7-11, the cbss P-, cannot have p-measure 1. Therefore, we obtain the foIIotnring coroIlarff

&= by schder W].

We remark here that the weaker statement that P- has p-meamre 0 wits proved by Schnler and Yamakami [98], and independently by Cai and Selman [Dl. Notice that, as an immediate comeqnence of

C o d h y 7.46, if NP E P-,,, then NP has pmeasure 0-

Deftnition 7.4.7 (Almost Immmity) [24, Tlj A set is d e d atmost P-immmre if it is a union of a

P set and a P-immune set. A set whose complement is almost P-immune is called P-h&k

For aset A, aset S is d e d amaxixdsubset of A ifany infinite subset of A in P is afinite variant

of S, te, IIS - All is Enhe. It is known that, for any h$nite reemsive set A, A is almost P-immune if aud o n I y i f A h a r a ~ s u b s e t inP [24,511.

hmma 74.8 [ 2 4 , q If a set A not in P s a t i q f k A se A via a length-inasaring redudion, then A is P-levelable. Hence, most well-known NP-complete sets m P - h a l e unless p = NP.

Proof. To arrive at a contradiction, we assume that A is almost P-immunee The aImost-P-immunitp

ensures that there is a maximal subset ofA in P. Let E be such aset.

Now wedefine B = {z I x E A f(z) E E). Smee f E FP and E E P, we condude that B is in

P. More~vef, B fl E = 0 by definition. We now show that B is inhite. Assume that B is hite. We

choose an dement x in A - E such that B zI~I-'. The element x exists because A # P bat E E P. Consider the set E, = {z} u {f ('I(=) I k 2 I), where f(')(z) = f (z) and f(k+l)(x) = f (flk)(z)) for each

k > 0. Smce f reduces A to A, E= A. Moreover, E, is h h k e and in P since f is length-increasing

(Le, [ f (2) 1 > 1~1). Note that E, n E # 0 since, otherwise, E U E, is a P-subset of A such that E - E. is

infinite,andonsequentiy EisnotamaximalsubsetofA Sincez#E, thereexistsastringy in Ez-E,

but f(g) E E. Hence, y E B. Clearly lyl > lzI. This contradicts the hitmess of B.

The Iatter part of the daim fo11ows from the fad that most known NP-cornpkte sets A, such as SAT,

Pw* 74.9 Assume that P # NP. If euery set in PPmmp - P is dmst P-immune, then NP g PP-cornp

R o o E Consider the NPampIete set SAT. If P # NP, then SAT is in NP - P and is P-levelable by

Lemma 7.48. Hence, if SAT E Pkmp, then Pkmp - P contakts a set which is not aImost P-immune

To show that an sets m P M ~ - P are almost P-immmre seems difEctxIt. Moreover, we do not know

how to construct a P-Ievelable set in PMmP which is not in P.

7.5 Closure Properties

This sectitm a9 be devoted to dosme properties under severd types of polynomiaLtime reductions- &eean

that, fix a reduu'bilitp I,, a class C is dtwd (dmmrtrrord) u& I,-rebrrcfionr it, fbr every two sets A and

B, A 5, B and B E C impIies A E C- We know that most k n m compIexity classes are dosed under pm-reductions, snch as RP, BPP, NP, PE, PP, @P, etc

In Subsection 75.1, we shall demonstrate that neither A:-, nor 22:-,,, k > 0, is dosed under p

m-reduciiilitp. The la& of this property indicates a structmal difference betmeen P-, and other regular

complexitg classes, such as P, NP and BPP. 'Rms, it presents a partial solution to the NP c?P-,, question. Oar resuit relies on the sets constructed in Section 7.3.1.

In Subsection 7.52, we We &ow that the dass A:-, is not dosed under the polynomially-bounded

operator. This d t dso implies that AEp-e~m~ is d X i t h m CE.

7.5.1 Polynomial Time Reduciiilities

In this subsection, we will discus properties of qnintessential complexity cIasses in relation to warious

polynomid-time reduciiiIities.

We cfaim that there is a the difkence between polyuomid-time truth-table redua'biIity and Turing

reducibility within Pkmp. This claim is p d by demonstrating that the Thing closure of PkOmp is

"large: whereas its truth-table dosare is "smalLn

The foIIowing proposition shows that the 'Rxing dosme of A~~,,, is equal to A T , where k > 0.

Ptvpodion 7.5.1 [94] For each k 2 1, the %ring eloswe of AEp,,, is equal to A T .

Proof. We shall show that every set m A? is pT-reducible to some set in AEPcomp. Let B be an

arb'- set in A; and assume that B E ATIME~(~, 2C") for some constant c > 0. Let A be the set d&ed

in the Hard Instance Lemma We then define the set H as

Smce A is in DTIME(n@"), clearly B' is m Af. Since B' G A, by the Hard Instance Lemma, B' is in

AA~-,,. Hence, P E A:-. Ftom flA n Zhl1 5 1, it Wows that, for all y of length n, y E B if and only if h(ln)y E B'. Recan

that h(ln) is computed by n adaptive qumb to Prefiz(A). This implies that B is pT-duciiIe to

ET $ Pre fiz(A). Let C = Bt Prefiz(A). Since B' E and Pre f k(A) E C is dso m

A~p-oorn, by Lemma 72.12. U

It is not known whether pT-redadbilitp m the above proposition can be replaced by pm-redua%ilitp.

Note that this is the case if Af;p-comP has some pm-complete sets for A;.

RooE Recall fhat EXP has prneisme L The coroIlarp then is immediate h m Corohy 7-46 since,

otherwise, EXP has pmeasare 0, a cmtxadictiam i3

Assmne that there exists a pmredction h m a set A to another set B which is in P*,. The next

ptopositicm dmazmke the reduction when A $€ P.

Proof. Assume that A is pm-reduciie to B in Pbmp via a Pampatable reduction fimction f.

Ebrthermore, assume that there exists a positive constant c satisfying log lzl 5 cl f (%)I for ahnost an z. In

the following, we want to show that A belongs to P.

Let MB be a &e which computes B in polynomial time on vd-average, and consider the following

machine MA: on input z, compute f (z), and accept the input exactly when MB accepts f (2). We thus

for some positive constant & S i B E P-,,, TimeMa,(f (z)) 5 2cIlf(r)l for some constant d > 0. This

implies that

TirneMB(f (2)) $ I f (=)I I ~ z [ ~ ' / ~

The next proposition, due to Wang and Bebger [6] (in a &t setting) and W e r and Yamakami

1981, shows the c l a m property of the class P-,, under restricted reductions.

Proof. Let B be any set in For a set A, we assmne that A 52 B via a P-compntable reduction

f which is increasing and phonest. We shall show that A is aiso in A:-,. For each p E P-comp, let

Y = prl. By the condition on f, we condude that v E Pcomp by Lemma 4.2.8. It is easy to see that if B

is computed by a deterministic ntriag machine M m polynomial time on v-average, then A is computable

in polynomial time on paverage by simdating M(f (2)). 0

As an example application of this proposition, consider the bounded halting prob1em BHP. Note that

BHP is pl-complete for N P under increasing, phonest reductions. Therefore, by Proposition 73.4, if

BHP E P-, then NP is included m PMp.

Next we show that neither A&,,,, nor I::-- is dosed under pm-reducibility. Note that Wang and

Belanger [Ill] have shown a similar result for their dasses APp, ANPp and DNP of distn'butional decision

problems.

Proof. We prove the theorem only for A;%, since the proof for C:* is almost identid Let A be the set in DTIME(O(n@* ")) such that llAnl = I far aU n E N and let h be the frmaion h computable in

time O(nb* ") in ordm to sepatate h m ATIME~(B, 0(2")), both of which are used in Theorem 7-39.

Now we define a reduetion function f as f o l k

~ o t e that if f(Pm) = 2, then n 5 a. ~ e t T = {di I i E N, f (d i ) E A). Clearly T 5% A via f. Note

that the fimction f is not phonest.

We shaII show that f is computable in time O(n). Assmne that z = 02i for some i E N, and let y be a

string of length [log 1zlJ2 in A. By our definition, the computation time off (z), where Izl= n, is at most

for any stafEdently large n.

We next show that T 4 Af;-,. A r n e otherwise. S i T is tally, T is in A:. Note that A! =

ATIME~(L, no(')). Consider the foIIowing algorithm:

begin input x

if z 4 A then reject

take n (5 m) such that f (a2") = z

ifPm ~Tthenaccept elsereject end.

This algorithm computes -4 and is in ~ T I l 4 E * ( k , 0 ( 2 ~ ) ) . This clearly gives a contradiction. Therefore,

T 6 A:P-m,- u

Theorem 7.5.5 yields the signiscaut consequence that the classes in the real poIynomiaI-time hierady

under P a m p are strndatralIy different h m most known woftt-case complexity dasses.

As another application of Theorem 7.5.5, we shall present the following coroIIary, pmven as Lemma

7.1 in [w. The corohy demonstratg the n e e d y of the phonesty condition in Condition I' d&ed in

Section rl7. The proof preseated here is very d&rent fiom Gmevich7s and is based on Theorem 7.55

Proof. Assmning the contrary of the corollary? we shall prove that PMp is dosed under pm-reductions.

This contradicts Theorem 7.5.5.

Assmne that A is pm-redtlcii1e to B via a reduction f, and dso B is in Ph,. We show that

A E Pmp. For every distribution p in P-comp, by our assumption, there is a distriiution v € P a m p

and a function p which is plynomid on paverage such that fi(v) 2 ZE # fm d y. Note that B is computable m polynomial time on waverage. Then A is also computable in polynomial time on paverage.

Since p is arbitrary, A bdonp to PP-comp- D

Let us recall Condition 1'. Condition I' asserts that every P-samplable distrriution is avpdominsted by

some P-computable distribution. Assaming Condition It, P-mp turns out to be dosed under mmy-one

redua%ility with phonest poLynomial-time computable reductions-

Proof. Immediate h m L e m r ~ 6.4.2 and by the same argument in Proposition 7.5.7. 0

E " d y we show the existence of an incomparabIe pair in Ph,, with respect to the h p m - r e d u ~ ' b ~

Proof. R e d h m Theorem 7.4.1 that there is a P-immune set in Pbmp. Let A be such a set, and

let B = {O}'. Note that A f! P. If A 52 B. then A E P since B E P, a contradiction- If B 5% A via a

pdynomiaIIy honest f E FP, then A should not be P-immtme since the bfhite set {f(z) 1 z E B), which

is a subset of A, belongs to P. Both cases mdnce contradictions. 0

Let us next obseme the closure property of average c o r n p I e cIasses under hpm-reductions when

P-sampIabIe distniutions are taken hstead. The P-samphb1e distniutions show us a mkrent worIcL

h o t We ody consider the case A!wp. Anmme that A is pm-redua%le to B via a phonest

reduction f, and B is in We sbdl &ow that A is in

Let p be any distxibuticm in P-samp and d&e v ap v = p p . Distriiution v may not be P-samphbIe,

but by Cornby 4.4J2, we caa find another d i s t r i e J which pdominates v and is in P-samp. Sina

B E A:-, we h a . (B,P') E Ava(A;, *). Note that (A,p) s& ( B , 4 via ff. Hence+ by Lemma

6-42(2), we ham (A,p) € Aver(AI, +)- Hence, A is m Agpl.mP- n

Hence, N is polpnomial-time bounded on yamrage. This contradicts our assumption. 0

Contrary to the sitmion for A:, it is not known whether A:-, E UP- A:pTomp, or ewn whether

A~p-~orn~ C, P. A:-. However, we are able to show that the converse does not hold for A;-,,.

Proof. Assume that Up-A:-,, A:Pcomp. By Theorem 732( l ) , we can define a set A in

m A ( k , 2n) and take a corresponding function h, as de6ned In the proof of Lemma 7.3.1, WE& has the

following property: for each length n, An E (h(In)) and h(ln) is computed in time 0(P). Now define

C = {On I 3y[lyl = n A y E A]). To see that C E UP- AEp-eom,, let B = ( ( ~ ~ ~ ~ , z ) ( 2 E A). By Lemma

7-52, B is also in AEp~om~, and let C = {On I 3y[lyl = n A (y,z) E B]). By our assumption, the set C is also in A:-,,. Smce C is tally, by Proposition 7217, C is in A;.

As a resalt, we have C E m ( k , 0(2")). Consider the following procedure that computes A:

This procedure gnarantees that A is in ATIMEb(k, 0(29), and hence, this contradicts the choice of A. O

-49 an immediate conse~nence d Theorem 7-5-13, we again have Pp-arrn,, # NP and PMP # UP. We note that it is not known whether P-AiFcmnp G CE- Howwerl if 3"-AE-, G CE-,,

then Theorem 75.13 i m p k that P- # NP-, and thus that P # NP.

Moreover* we can consider a restricted tlgpe of probabilistic operator.

Dehition 7-5.15 (Class PP- C) For a complexity dass C, a set S is in P P - C, the d m under Mc

prrrkbibtic opaatm, if there are a polynomial p and a set B E C such that S = { x I ll{y E P(lf 1) I (y, 3 E B)ll > 5 - 2flz"}-

The p r o b a b i i c daPs PP is dosed under this operator, ia, PP- PP PP. A proof techniqne similar to that used for Theorem 75-13 shows the foIIowing theorem.

Proof. Let the set A and the hction h be as defined in the proof of Theorem 1-5-13 but with k = 1.

Recall the algorithm descriied in the proof of Lemma 73.1. Define B to be a set accepted by the foUowing

algorithm

begin input x (say, 1x1 = n)

compute zr = min{w E I w $! KT[n, ~(0 '" ] )

i f l ~ r ~ t h e n y = O e b r e y = l

if z E y ~ l * l - l then accept else reject

end.

By the choice of y in the algorithm, A f7 B = 0 and llB fl PI1 = 5 - 2" for an n. Note that B is in P. Now m Iet A' = A U B. Then, re bave A' E Pkmp Note that either I1A' f l PI1 > $ 2" or IIA' n PI1 = f - 2".

Let C = {On I Il{y E P I y E A')ll > i - 2 )̂. We have C E PP- PkmP by definition, but C 6 Pkm, since if C E Pkmp7 then the same aIgorithm as m the proof of Theorem 7.5.13 computes A m time 0(2^).

0

Theorem 7-5-16 directly shows that PpzomP # PP.

7.6 Bounded Error Probabilistic Polynomial Time

We direct our attention now to the bounded-emr probabilistic complexity dass BPPr- In this section, we

shall discuss the dass BPP3 mrda weak reductions, and consider a result d Schnler and Watanabe [96]

regarding the qnestion NP C-7BPP3. We begin with the dosure property d BPPp,, under phonest bpptt-reductions-

PmaE Assume that A E BPP- and A is bpptt-reducible to B via a phonest reductioa Let

CHAPTER7. QUINTESSENTIAL C O M P m I L l T Y 267

M be such a reduction, ie., a polynomial-time probabilistic 'Ihring machine computing A with bounded

error which accesses oracle B with nonadaptive p h o n e gueries. For simplicity, assume that the number

of queries on each computation path is of the form 2", where m is dependent only on z. Since M makes

phonest queries, for any query z by M on x, 1x1 is bounded above by an abso1ute polynomial p m [zl.

Choose any distribution p m P-samp. We define a distribution v as foIlows:

It is not difEdt to confirm that v is P-samplab1e. To see this, let Mp be a sampling machine which sampies

p and consider the following sampling algorithm:

begin sampling algorithm

input 0'

simulate M, on 0'

let z be an output of Mp on 0'

list a s y z19z2r---,z2m

generate k (I 5 k Zrn) at random

0-a Zk

end.

Hence, (A, v ) E Aver(BPP, P-samp). We now show that (B, p) (AT Y ) via M . It is sttfticient to check

the domination condition for M. This is easily d e d , however- By Proposition 5.5.7(4), (BT p) beIongs to

Aver(BPP, P -mp) . 0

Ben-David d d [9] showed that all distri'butional NP search prob1ems are Ltaadornly" reducible to their

corresponding distniutional decision problems. Later Scbnler and Watanabe [96] rephrased this resalt in

the foflowing f&ihion.

Proof. Assume that NP BPP-,,. For any set A E g, there exists a set B E NP such that

A is pa-reduuile to B. This reduction can be phonest by choosing an appropriate set B. Hence, A is

bpptt-reducii1e to B via some phonest reducti011, Notice that B below to BPPWp by om assnmptioioe

By Proposition 7.6.1, B E BPPWp implies A E BPPP--~. Theretore, we get A E BPPwp.

Impagliazu, and k i n [q demonstrated that distributiod N P decision probiems are "reduaLbIen to

distributiod NP decision problems with standard distribution. Schnter and Watanabe [96j extended their

resuIt m terms of- compntabiE@

Proof, We follow the argument given by Schuler and Watanabe [96]. Let us asstme that NP E BPPw~~sorn,.

Let (A,p) be an arbitmy distn'butional decision problem from Dist(NP,wP-samp). We shall show that (A, p) is m Aver(BPP, *). If B(C) > 1/2 for some n E N, then we can diminnte an strings of Iength

n from the following argmned- Hence, we annnne without loss of gen* that ji(Zn) 5 1/2 for dl n E N S i A E NP, there are a set E E P and a po1ynomiaI p such that A = {x I 3y E P(frl)[(z, y) E 4).

Moreover, h m the fact that A E EXP, we can dehe a deterministic 'Ihring machine MA which recognizes

D m time 2dlrl), where x is any input. For the sake of convenience, we ansume that ~ ( n ) 2 32n3(n + I)* for

d n € N

Smce p is an average P-samplable distribution, there are a generator M for p and a po1ynomial p' such

that (i) Iji(z) - P~~[M(O') = zjl 5 r', and (ii) for every r > 0 and n E N,

Write pj(n) for p'(2j+'(n + q(n))). Let us dehe the function f as ~OIIOWS: for each j with 0 j < q(n),

Therefore, we obtain

For each 2, since LSd,,) 5 1, there exists an index j' such that j(z) 5 Hn) + & - 23*(")+ee'. Let jl be the minimnm of such an index.

On the other hand, if q(n) < p,- (n) - &) - ilog(q(n)), then fi(z) 5 2 - 2-dn) = 2-dn)+I.

Let x?) = (2 E C" I j = j, A q(n) 2 pj=(n) - &) - ilog(q(n))). It foUows that

since q(n) 2 2n3. Therefore, fi(x2)) < $ 2-J for all j with 0 j < p(n).

RecaII that sn denotes the nth string of C' (NB. X is the 0th string), and s p represents the kth string

of Pw("') (N.B. S? = o~'%(~)). Note that Isn( = llog(n).

We begin with the fo11owing NP sets:

A hz of? o hl(v) = h2 of,?' o ht(u1) = =]I;

where I = m - k, h~ E HI,,, and h2 E H4+2- The notation h14z above meaus the concatenadon of three

strings At, ha, and z. Let B = Vi=, Bi. CIearly B is in NP. By the assumption, (B, v-) E Aver(BPP, *)-

Let Q be a d a e n t l y Iarge and fbred amstant* First we s h d dehe the randomized ?tuing m;lrhinp

MO that (i) chacse j at random finm {0, 1,. . . , q(n) - 1) (by choosing a?)), (ii) chooses k at m d o m

h m {O, 1,. . - ,pj(n) - 1) (bg &-g @')), a d (E) (z, ~ f ' ~ ~ ' @ ~ ~ I)). The fohwhg ir a f d descdptimafthisaIgorithm:

begin randomized agorithm for Mo

mpat r (my, = 14)

-Pe ilog(q(n)) generate at random a string s of length iIog(q(n)) such that

s = s ~ ) f i x s o m e j w i t h O ~ j<q(n)

(asllme that s = SF)) -Pe ilog(pi(n)) generate at random a string d of length iIog(pi(n)) such that

The machine Mi rum in poIynomial time because the running time is proportional to the Iength of the

random seeds. The ten& of random seed, is^(")@')[, is at most

for some constad d > 0.

Write d(n7 j7 k ) = s:(~)$(~). La us define h(z) = d(nlj,, &I) and D = ((z, h(r)) I z E Lw ). Note

that, for any z and y, if z , y E x?', then Ih(z)[ = Ih(y)l. For alI z E x,O'Il since Ih(z)l 5 c'(j + Iogn),

we have Ih(z)[ I d -log(l/n2 fi(x2')). By Lemma 3.3.11(3), we condude that k.lh(z)l is Io@thie on

Pa-

Next we d- another randomized orade Turing h e N as follows:

on input (2, h(x)) is at most

for some appropriate constants E', d > 0. T~US, for each z E x!?, d

1 ~ s ~ m e f : ( ( z , h ( ~ ) ) ; s ) I s E r ~ ~ ( ( ~ t h ( z ) ) ) ] < d * -I- d'c'. (a(x .D))n2) -k c'.

By Lemma 3.330, we condude that k .ES[r rme~( ( z7 h(z)); s) I s E rN. ((2, h(x)))] is polynomial on p

average. Therefore, hsJVB((z , h(x)); s ) is polynomial on p-avefage.

Next we shall discuss the success probability of machine N. Fnr z and let n = 121. Let p, be the

probability +dm the algorithm N on input (z, h(x)) succeeds in producing the conect amver. We shall show

a Iower bound of p,. For this purpose, we introduce a new notion. We say that (hl, h2, z) d e k m h s z if

(i) there Bdsts a string v E Zl such that 0 hl(v) = z and h2(4 = z; and

(ii) for all lo E !St, if h2 o fi) o hl (M) = t, then f.) 0 hl (w) = Z.

Fa j and assume that 11lf$))-'(z)11 # 0. Let us chose @ and set I = pj(n) - &'I. Then, p, 2 1/16

fdm fiom the claim betow.

Next we rhall show that n2(hI) 8. Let FZaI = ran( f:) o h l ) The cardinalitp of FZvht b at most 2.

To complete the pmof of the claim, we combine the above results.

This completes the proof. I

Finally we apply Lemma 55.3, and thus there eJdsts a randomized lS.trjng machine MI such that

hs& ((z, h(z)); s) is polynomial on paverage with the condition that, for each s' E l?~,(z),

By the remark fo1loPPing Proposition 3.53, we immediately obtain that (A,p) E Aver(BPP, *). The proof

is completed. 0

The class BPPhmp is d d y related to P-samplable dishiiutions. In the following Ie~nma, we see

that an assumption like NP BPPPeomp causes the P-sarnplable distriitttions relative to NP oracles to

be average P-samplabk

Proof. We shall show only claim (2). Assmne that E BPPpamP- Take an a r b i i ~ ~ ~ - s a m p l a b l e

distriitttion ~ r . By its definition, there exists a sampling a1gorith.n M, a polynomial q, and a set A E NP suck that, for any string z and any number i E N,

Iji(z) - R ~ [ M ~ ( o ' ) = z m time q(lxl, i ) ]] 2-i.

We can assmne without loss of generality that if M wi th orade A on input 0' haIts and outputs z, then its

nmning time does not exceed q(Ix1,i)-

Let us define the set B as fdm

B={(0',ln,s,1j,d) Id€ {O,I}A~E~~~~~AIM~(O~;S)~ =n

A Tiie&(oi; s) 5 q(n, i ) A the j-th bit of MA(Oi; s) is d ).

Let p(n) = 2n+4 for dl n E N Smce B E bP, our assumption ensures the existence of a probabilistic 'king

machine N which accepts B m polynomial time on Y--average w i t h error probability 2*n)-n-'4, where

input is of the form (O', la, s, lj, d). Consider the fbIIowing sampling marhine W.

generate $ at random so that lss'ls q(n, i + p(n))

sets=ss'

for j = 1 t on do

if N((o~+P(~) , in, S, ij, 0)) = ~ ( ( o ' f d ~ ) , in, S, ij, 1)) then go to (*)

if N((o'+~(~), in, S, IJ, 0)) < N ( ( o ~ + ~ ( ~ ) , in, S, ij, 1))

thensetdj=lelsesetdj:=O

end for

output dl& - - - d, and halt

(*I end for end.

In the foIlowing analysis, we fix i and z. Let lzl= n. Set

Write o', to mean P~,[M~(o'+P(~)) = z in time q(lz1, i + p(lz1)) I s € RMA (o'+~"))]- By the dehition of

St,i, we have o', = CaESmSi 2-la[.

Note that the success probab- tfmt each iteration of the second for-loop is at least.

Now let us denote by p i the probabiIity that the machine M' on input o'+P(") with random input s E S=,i

outputs z. The fimction p i dearly does not exceed o',. The Iower bound of p i is dcnlated as fo11-

The last inegnality folIows from Lemma k6-

We then have

As in Theorem 4-4-13, we can show that EJ - p i ] 5 2-' + 2-j for all i, j > 0.

It is not diflidt to show that N' nms in polynomial time on paverage with respect to the size of its

output. Thedore, p belongs to avP-samp. 0

Proof- Assume that NP E BPPhmp. Notice that PC-samp E P avP-samp implies BPPmpamp

BPPpy,. By Proposition 7.6.4, it foIIm that NP E BPPmwp BPPP2~-runp- A simiIar argtxnent Ieads to (2). 0

7.7 Random Oracle Separations

We return to Levin's original qnestion of whether NP E Pkmp. Since this question is difficult to answer,

on. we turn our interest to its reIativizati

Bennett and Gill [8] 6nt d a notion of 'andom odes" and made several important mntriibutious to

computational complerity theory induding the r d t that, dative to a random oracIe, three dasses P, NP, and mNP are Merent. Intrritnety, if we choose an orade set A at random, PA is difberent h m & with

probability 1. In other mads, "most" oracles can separate P h r n NP. In the same paper, Bennett and

GiIl proposed a 'tandom orade hypothesis" that states if a property P holds relative to a random o d e ,

then P dso holds m the non-rektivized world- (Hmwer, this hypothesis is now known to be fdse [a*) To approach Levin's original qnestioq we consider randomly mhhized world We remark that there

is no known inclusion relationship between the dass Pbmp and other wo- complexity dasseq such

as NP and PSPACE. This d o n wi l l discus, reIative to a random oracle, several negatbe resnltts about

these relationships

We identifir a set A with a binarg red nmnber 03 , where r = ~ ( 0 ) ~ ~ ( 1 ) ~ ~ ( 0 0 ) ~ ~ ( 0 1 ) '- -.

De5ition 7.7.1 (Random OracIes) Let @ be a property r e h b i d to o d e X. We say that, (with

p r o ~ l ) ex ~ o ~ r d o t i ~ c t o a ~ m ~ e t ~ ~ t h e ~ b e s g n e m e a s l r r e o f t h e s e t {XI @hoMs}is

1, denoted by m((X I @ halds }) = 1.

In om setting? the Lebesgne meamre behaves Iike a probability measme if the property is W y

evaluated. Here is an emmpIe. Let

Then, cI&y Iv~(z)l = lzl for all z. Consider the event (A I m(z) = o ~ I ) for each 6xed string z. The

measnre of this event, m((A I q ~ ( z ) = Olfl)), is 2-lrI. Hence, for each n E N, m({A I Vy(qA(y) # On))) =

(1 - 2-n)2m, which approaches l/e as n goes to oo.

We shaU show that, with probability 1, NP-, is different from P-,, relative to a random o d e -

In this section, we choose the fdowing relativization of the classes PF and NPF.

&hition 7-72 (Refativization) Let X h a set of strings, and Iet F~ be a set of distributions relative

to X.

1. Let PgX be the collection of all sets A such that, for any distribution p in F ~ , (-4,p) belongs to

~ v e r ( P , F ) ( ~ + ) for some distribution v.

2. Let NP& be the collection of aU sets A such that, for any distribution p in TX, [A,p) Mongs to

AW(NP, 7)(X~) for some distriiution v.

Proof. We can even show the slightIy stronger statmmt that TAUY n NPX CO-NP~ relative to a

random oracle X. This clearly yields the desired conse~nence. In what follows, we shaII prove this stronger

statement.

We ht inwdttce a test language which Lies in I'ALLY n @ for any oracIe set X. For this p-

let ns d&e RANGEA = {On I 3y[m(y) = On]). CleatIy RANGEA is in W Y and also in @ for any

oracle A. Let M be an arbitrary polynomial-time o d e Tming machine Let CO, = {A I Vy[q,&) # On])

and let Ck = {A I m(On) # On h 3 ! y ( ~ ( y ) = On)]. As seen above, m ( c ) = (1 - 2-n)2n for any natural

number n, and c~~l~equently b, m(Pa = 11% where e is the base of the natural logarithms.

Let Y be the probability space of C - {On). We introdace a transformation from CO, x Y to C: x Y as

f (A, y) = (&q~(y)), where A, = A - {ylOm 1 1 I rn I n}. We now claim that f is For any

(B,u) f Ck x Y, there exists a unique y such that I&) = On- Let A = B u (ylOm I 1 5 m 5 n). Then,

q ~ ( 4 # On for an 2 because ~ A ( Y ) # On- Moreaver, for any event E G Ck, we have

Hence, in particular, it holds that

Let as de6ne two conditional probabiIities Let

Proof of CIoim Let n be large enough that 5 &. For any pair (A,z) E CO, x Y, with conditional

probabi i ao, MA accepts On. Let ns fix one of its accepting paths and denote it by p. Let us consider the

oracle set A=. On the path p, the probabii that M does not make aI1 queries of the form zlok, 1 < k < n, is at least 1 - 2 g. With this probability, MA* accepts On. Thus, (lr, can be estimated as f o h :

99 m((A I On E L(M, A) A A E c}) 3 - - loo m(Co,)

FinaIl~ we must c a l c u k the overall error probability c = m({A I MA(on) # RANGEA(On)}). Choose

an integer n Iarge enough that we can guarantee that m ( c ) > $ and 4 > $4. For this n,

Hence, the ermr probability E is not 0. Th, the event that lldA computes the complement of RANGEA has measure 0 by the 0 - I Law. In other words, RANGEX 4 c&Ip relative to a random oracle X.

Using the above proposition, we can show the following separation d.

Proof. For any oracle A, we define RANGGA as the set {z I 3y[q~(y) = 4). CIearIy RANG&A is in

w. In the following, we want to prove that RANGGx is Px-bi-immune relative to a random orade X.

First we claim that is Px-imwure relative to a random oracle X. Let us 6x a polynomial-time

deterministic oracle Thing &e M and I d p be an increasing poIynod whic6 bounds the running

time of M. If L(M, X) is finite reIative to a random oracle X, then the ciaim is trivially true. Now we

assmne otherwise.

We say that y is e x m h d by M on input w if MA on input w queries a string of the form ylOm,

1 l m l l ~ l * Let

EXAMA(w) = {y I MA on w examine y, but there is no v < w such that M* on v examines ).

Using the set EXAMA(w), we dehe W I D A as foUows:

Now we shall show that, relative to a random oracle X, EVIDX is W e .

Proof of CIdim Note that, by the polynomial bound on M, IIExAM~(w)~~ 5 p(lut1) for any string w. Now

fix any sufficiently large string w satisfping IIEXAM~(W)~~ I 21w1/2 for all oracles A. For each y,

For simplidty, we write FA(w) to mean that there exist two sttings z and y such that x 2 w, y in

E X M A (w), and q ~ ( y ) = zrz. The probability that FA(w) holds is at most

Hence, C, =({A I FA(w) holds )) 5 291f < oo. Thns, by BoreK:adeniTs Lemrna (Lemma A.11), m ( { ~ 1: WFA(W))) = 0. ~ h i g implies that m ( { ~ I WIP is idbite 1) = 0. I

Let us define D = { A I L(M, A) is an in&& subset of RANG! and E V I P is bite ). To Iead to a

contradiction, it d c e s to show that m(D) = 0. To do so, for each set A, we shaII define the series of strings

{@)ia as f d m initiall9; set 4 = A, and for each k 2 1, set

Note that zf may not always be dehed. Using this sedes, we de6ne

Dk = {A I zf exisk and Vi c k[zf E RANG:])

forkzl . It isnotdifEcult toseethat Dl 2 D 2 2 - - - > D .

We next daim that, for almost d k, the ratio m(&+l)/m(&) does not exceed some constant c less

than 1.

VP U E E X A M ~ ( @ [ T A ( ~ # z=l A 3 6 U EXM~(W)[~A(Y) = -1; e w l z

Smce = DHI 17 4, the ratio m(&+~)/m(Dk) is e q d to the conditional probabiIity m(Dr+* 1 4). This is bounded by the probability that there exists a shing x accepted by M with o d e A such that

q ~ ( y ) # 222 for dl 9 in E X A M A ( w ) for some zu z, but ~ ( f ) = m for some y' which is not in

Ue EXAM*(W), provided that $ exists. Hence,

The last expression tends to 1 - 1/e as k approaches oa a

Therefore, for almost all k, rn(D~+l) c m(Dk), which implies m(D) = Limk, & This yields the - conciusion m(D) = 0.

The proof that RANGqx is pX-immune relative to a random o d e proceeds simikrly. 0

h o t By Proposition 7-75, there exists a ~~-bi-immtme set in for a random oracle X. Smce Proposition 7-45 is relatbidIe, Pghrnp ha9 no PA-bi-immune sets for any oracle A. Therefore, D@

earmot be induded in ex,, dative to a random o d e X. 0

Take any polynomial-space de* ' ' tic Turing machine M. For each z, Iet C, = {A I ~ ( 0 2 ) @

Q(M, A, z)) and = {A I A $! C,). We shaU show that m(C,) is not smaU

Crcrim 22 m(C,) 2 1/3 for almost all sbings z.

hf of C k h . Notice that the tape space used by the h e M on input z is bounded above by a poIynomial m lzl. Let p be such a poIynomiaL By the space b o d , the wmber of all possible instantaneous

descriptions of M on input x is at most 2~(bl), and this nmaber is irreIevant to the choice of oracles. This

number is aIso an upper bound of the number of queries made by M. On the other hand, to determine

m ( P ) , we need to check at most 2M-' b i i since

The measure of c, is at most the ratio of the number of query strings to the number of bii to determine

ObvionsIy, m(EJ approaches 0 when z grows, and therefore lim,, m(C,) = 1. I

Next let us consider the error probability c = =((A [ ~ ~ ( 2 ) # QUERY(z))). We must show that

e 2 5. Note that

To obtain the desired lower bound of c, we d&e two sets

Next we shall show that m ( c ) = m(Cf). To show this, we f f i e the transformation f, as

It is easy to see that f, is bijective on C,. Hence, from the fact that m({A E Cz

m ( c u Cf), it follows that m ( c ) = $ m({A E C, I z f L(M, A))). A similar argument shows that m({A € C, I z 4 L(M,A) A r)~(O'f) E A)) = $

L(M, A))). By combining the two eqnations, we obtain .

Chapter 8

Conclusion

In the early days of average-case analysis, d e r s had great hopes of solving aIl NP-compIete problems

in average po1ynomial time. The satisbbility probIem, for example, can be solved m average polynomial

time with respect to some n;rtnral input clistrilution, and so have the Hdtonian circuit prob1e.m with some

input dishiiution. As resear& has progressed, however, the realm of averagecase analysis has encountered

the same adty as its worst-case comteqart.

L Levin showed that there are pmb1ems intractable even in the average-ose setting, Since Levin pre-

sented his approach to average-case c o m p I e theory, much research has bee^ devoted to grasping what

the intractability of problems is. These a- adyses have attracted s&nScant attention h m re-

searchers in the other Ed&, such as cryptography and statistical physics- Researchers have continued to

seek for another axnpIete prob1em for Dist(NP,P-comp). At the same time, there have been a number of

diikmnt approaches dedoped to gain a better understanding of a- hfmduble probllm- One of

them is to study the a- c o m p I e of distrriutional search prob1ems. For example, NP search

prob1err1~ have recedy been shown not to be harder than NP-probIems in the a- setting.

This thesis tried to mtri'bute to the development of a general and consistent theory of a m q p a e

c o m p I e - PersonallyT I have been insp'rred by Levin's early question of whether atl NP-compIete problems

are soIvzb1e in poiynomid time on the average with respect to aatnrany selected distrributiom. This westion

is deeply related to the E =?NE question as well as the P =?NP question in worst-case comp1exity theory-

We thns see a tie between amapxse complexity theory and worst-mse compIexity theorp,

Reduu%ilities have phyed a eentrd role in our stndy of the stmctntal properties of those dases. One of

the significant f&tures ofthis thesis is the introduction of two a- versions of the po1ynomiaI-time

hi-, oneby a ~ ~ g r e d u a % ~ e s , and the other by amodel of a1tern;itingTbbgmaches

These two hierarchies pteser~e nmnerotrs properties of the worst-case hierarchy, but they are idrinsicaIIy

difkmnt in c k a c k r because of their sen- to the choice of input distn'bntins, A stntcttllral study

of average classes has commertced in recent years, whexeas there have W y been a namber ofstndies on

worst-case cornpIexi@ dasses.

Another important featnre of this thesis is to introdnce the notion of comp1exity dasses.

-8. CONCLUSION 282

This notion actually enables us to bridge the gap between a- complexity theory and its worst-case

comterpart. In particular, the class P-,, has been thorougidy studied in this thesis. We have seen a

variety of possib'ilities in this direction that contri'bute to an understanding of averakpcase comp1exity.

We regret that there remain many intriguing topics in aveTage-case c o m p I e theory, and we have left

numerous qyestions rmso1ved in tbis thesis- One important direction is the study of distri'butionat search

problems. There are a number of studies in this area (see ag, [9,12, 113,1081).

Average-case complexity theory is a fraitfnlfidd to dtivate. HO-, it is not always easy to seek

the right definition of a- counterparts of &known concepts in worst-case compIdty theory.

For exampIe, we have aIready seen several possible ways of defining the class Aver(NP,F). Consider

the follwing exampIe. Suppose that we wish to ask otuseIves "Is it posslile to dehe in a natwul roay

an average-case version of #P ?" Recall that the #P-frmctions were originaiIy introduced by the model

of polynomial-time counting machines (LC, nondeterministic T b h g machines which sum the number of

accepting co~lfiigmatio). Our definition of nondete ' 'stic ntring machines does not seem to provide

any reasonabIe model for counting ma&ines because not all accepting computation paths are considered in

measuring the nmning time of the machines. One possible way to dehe such a dass, say Aver(#P, F), is

as fonow~:

IMnXon 8.0.8 (Average #P hctiops) A frmction f on E' is caIIed #P-mnpz&le on p-wenye

if there exists a randomized ?hhg machine M such that (i) M is poIynomia1-time bounded on p-average,

and (in) f(z) is the number of accepting computation paths of M on input z- The class Aver(#P,F) is d&ed to be the mnection of an pairs (f, p) such that f is a kct ion which is #P-cornptrtabIe on p-average.

R Impagkzo recentlp expressed the view in [43j that there are five possiibIe worlds we might inhabii: AZ- 9-ca, in which no fntractabIe probIems exist (i-C, P = NP or N P BPP holds); H- in which

there are intractable pmbIexns, but no problems are hard on the average (ie, P # NP P-); Pessi- in which there are hard distn%tttional probIems, but no (strong) one-way functions exist; Minierypt, m

which oneway fimctions exist, but public key cryptography is impossiiIq and C~yphmtk, in which pubic

1Qy crppWPphs is pclssl%le- No matter which world we researchers w d l contime to pursue attempts to understand natural

phenomena.

Appendix A

Small Lemmas

This section provides several important Iemmas used in this thesis

The following inequality is known as Matkovrs bqdhy.

h m m a A . 1 (Markav's Ineqnality) Lefpkadiskibutiunandlet f b e a J h c t h f r o m C ' to* .

For every positive d nu* r, ji({z I f (2) > r - EV(X)l}) < 5, where EV(X)] = C, f (z)ji(z) .

Proof- If EV(X)] = 0, then either f (z) = 0 or P(x) = 0 for all z E C'. Thus, if f(z) > 0, then ji(z) must be 0. This yields the amsequence that ji({z [ f (z) > 0)) = 0 < 5 for all r > 0.

Now we assmne that EV(X)] > 0. Let A = {z I f (z) > r - EV(X)]). Since the case A = 0 is trivial, we

assume otherwise. Dehe tf (z) = 1 if f (z) > t - EV(X)], and 0 otherwise. S i tf (r) < -*11 for dl

ProoE In the case where E[XI = oa, we get

fordlpositivereal nmnbersu, v, and w satisfyingu~ v cut.

Let us fhc o and consider the upper and lower - By the monotonicity of the fanction f , these limits g*(v) and g.(v) exist for any u in the interval (0, oo).

The definition also implies g*(v) 2 g. (v).

Proorf of Claim If z = u, then the claim is trmiaI. Now suppose z > v. Then, by (*), it follm that

f (4 - f (4 f (4 - f ( 4 - 2 - V 4 z - v - 9.(4 < .-

Hence, f(z) - f (v) I c - ( z - o ) .

In the other case where z < v, we get

and thns, we condude that f ( z ) - f (v) 5 c . ( z - o). I

As a s p d case of the above claim, for a random variable X, f ( X ) < c (X - E [ q ) + f (ELXI). By

taking expectati-,

Froof. We &st show that (I) implies (2). This is somewhat straightforward.

k(z, 0') - g(z, @)I = )g(z, 0') - h(z) + h(x) - g(z, @)I

C o d y , we show that (2) implies (1). Asgmne that Ig(z,Oi) - g(z,Di)l 5 2-' + 2-j for almost dl

i,i E N We first note that, for each 2, there exists the limit hj- g(z, 0') by a classical argument m

analysis. We Keen set h(z) = lim- g(z, 0').

Then,

Proof. By induction on n, we shall show that

where nE: lei = 1 if k = 1, aud [bj1 = 1 if k = no The lemma immediately fo11ows ham this

ineqtraIity.

The base case n = 1 is tmia. Assume that n 2 1. S i laA - bB[ 5 la1 - IA - BI + 10 - bl - IB[ holds in

generat, we have

Proot The proof proceeds by induction oa m In the base cases that rn E {O, I), the daim is trivial So,

we m e rn 2 t The induction hypothesis says that (1 - ?-f)'D-l 2 1 - F*-*. Then, we have

APPENDIX A. SMALL LEMMAS

Proof. Suppose 2(k + 1) I n. Notice tbat this assumption implies that 2k + 1 < n since k and n are integers.

Therefore7 we conclude that > ST;.

BeIm we shafI state Stirhg?s form& without pmk.

Lemma A. 8 (Stirling's Formala) n! = =(:)"(I + & + h(n)), where h E 0($) rmd e is the h e

ofthenafTd&7g-

ProoE Suppose that n is even and is of the form 2m. Using Stirling's fonrmla, for any snfficiently Iarge

ThelastinequaIi~foIlowsfromthefact that l+& < &. Fromthefaathat (I+&)" 5 &?it fo1Iows

that

Lemma A. 10 Let I be a finite indQ set. Let {&)iEr be a partition of the sample space $2. Then, for m y

ProoE For simplicity, assume that I = {1,2,. . . , k). By the definition of the conditional probabihy,

Pr[& I &] = R[& n &]/PrE, [ I if R[&] > 0. Hence, we get Pr [& I &] Pr[&j = Pr[& n &I.

Sice Ei and Ej are disjoint if i # j, the sum of all Pr[& I &] - R[&] is calculated as foIlm:

Given a series {A+)iEn of sets, the notation b s n p n A, denotes the set c, rh -4c and is called the

Iimits of {&)ia. Note that w E I.imsttpn A,, if and only if w lies in inhibdy many of the &.

PmoE (1) Assmne that &[A,,] ammge% Let k be anp natural mnnber- h the fact that

limsnp,A,~U"i=LAi,itfonowsthat

By om assumption, R[&] = 0. Therefore, Rmsap, A,,] = 0.

(2) Assume that {&lie is an independent s e q ~ a ~ e of events. Assame also that Pr[A+] diverges.

It d c e s to prove that R N z n E AL;] = 0, where AT is the complement of At (i.e., f2 - A&). For this

dsLo,aewant toshonthatR[n~A~]=Ofordn~N n+j c Note that P r m k A;] = hi- Prn, AJ. Let e be the base of the natural logarithm. Since

1 - z 5 e-' holds,

where erp[t] meam i. Let us consider the last expresion and denote it by Ti. Since RL4i] diverges,

the value Ili, tends to 0 as j approaches infinity. Thos, Iim j- A:] = 0. 0

Bibliography

[I] E.W. Mender and R Rubinstein, P-printable sets, SL4M 3. CompL, 17 (1988), 1193-1202.

[2] D. An&in, On counting problems and the poIynomial hierarchy? !l%eorrtical: Computer Science 12 (1980).

pp.161-173.

[3] T. Baker, G. GiII, and R Solovay, Rekhbtions ofthe P=?NP question, SLAM J. Cornput. 4 (1975), 431-442.

[4] J. L. B a l k , J. Diaz and J. Gabarr6, Struhaol Cmnplcziry I, Q Springer-Verhg, 19880, 1990@).

[6] J. Belanger and J. Wan& Rankable diskiiatiom do not provide harder instances than anifann dkhhtious,

in Ist Computing and Combimatorics Conference," 1995, Lecture Notes in b p n t e r Sden~e,

VoL959, pp.41049.

[fl J. Beknget and J. Wang, Reductions and con-ae rates of average time, m "Roc. 2nd International

Computing and Combinatorics Conference", Lecture Notes m Computer Sdence, VoL1090, p p ~ 3 0 9 , 1 ! W .

[8] C.H. Bennett and J. Gi& Relatme to a random orade A, # @ # u+w with probablitp 1, SUM J.

Comprt. 10 (1981), 96-113-

[9] S. Ben-David, 13. Char, 0. Goldreich, and M, Lnby, On the theory of average case complerdtp, J. Compct Sysfcm S d 44 (1992), 193-2l9. A prdimkprg oersion appeared m Pmcdbgq 22nd STOC, 1990, pp.37%386.

[lo] L Berman and J. Haraoaais, On isomorphism and density of NP and other complete setq SLAM J- Campat 6 (1977), 305-322.

BIBLIOGRAPHY

R Book and D.2. Do, The a;isteaee and density of generalized c o m p w cares, J. of ACM, 34 (1981),

7lS-730. 1987.

R Book, DZ. Dtr, and D. Rrsso, On polpnamial and generalized complexity cores, in "Proceedings, 3rd

Conference on Structore in CampIadty Thmf, pp236-250,1988.

J. Cai and L A. H e d r a , On the power of parity polynomial time, Moth, Systems Ttreory, 23 (1990),

pp.95-106.

J. Cai and A. Selman, Fme sepamtion of average time c o m p l e dassg, in Tmxd&s, 13th A n n d

Symposium on T b m t i d Computer Sciencen, Leaare Notes in Computer Science, VoI.1046, pp.331-343,

1996.

J. Carter and M. Wegman, Universal dassg of hash frmctions, J. Comput. System Sci., 18 (1979), pp.143454.

A.K. Chaadra, D.C. Kozcn, and LJ. Stodrmeyer, Alternation, J. ACM 28 (1981), 114433.

S. Cook, The complexity of theorem proving procedures, in "Proceedings, 3rd ACM Symposium on Theory of

c ~ m p ~ k g , pp.15i-i5s, an.

K- de W, E-F- Moore, C.E. Shannon, and N. Shapiro, Computability by probabitistic machine, in C-E.

Shannon and J. McCiuthy, e d b s , Automotn Sttrdiy pp.183-212, Princeton University Ress, Princeton, 1955.

DX- h, T- hkowhz, and D A Rum, Structural properties of complexity cores, unpublished manusaipt,

Department of Mathematis, Unbmky of Catifornia at Santa Barbara, 1984.

J. E h c o and M. Paull, Probabilistic andysis of the Davis Putnum procedure for solving the satidability

problem, Discrete Applied Math 5 (1983), 77-87.

M a G a ~ p and DJ. J o k n , Comprtss ond I-, A Guide to the Thtmy of NP-Compfetenus,

W E F h m a n and Company, Ne-a York, 1979.

J. Geske, D. Huynh, and J. A note on a l m ~ h e r e complex sets with application to polyaomial

compIexity degrees, Infmrnorion ond Comprdation 93 (1991), pp.97-104,

J. Gill, Compatationa oompkity of pmbabilirCic nPing machines, S W J. Comput. 6 (l9n), 675-695.

M. GoIdma~, P- &ape, and JJhbd, (m average time hierarchy, Infmm. lhxs+tg tctkrs 49 (1994),

15-20.

0. GoIdmich, Towards a t6eorp of amage case compIeJdtp (a survey), Technical N0.507, W lnstitnte

ofTechaology,=

J. Goldsmith, L.A. D. Joseph, and P. Young, Near-testable sets, SljlM. J. Cornput 20 (1991),

pp.506-523.

P. Grape, Taro results in average case complerdty, Technical Repat TFU'I!A-NA-9105, RoyaI Institate of

Technolagy,SStodrhokn,Saeden, f99L

BIBLIOGRAPHY 291

[34] Y. GafeYich, CrmrpIete and incomplete randomized NP problems, in Tmxdb@, 28th IEEE SymP0s;am on Foandati011~ of Computer Sdence", pp.111-117, 1987-

[35] Y. G d c b , Matrix decomposition is complete for the average case, in *Roceedings, 3lst EEE Symposium

on Foundations of Computer Science", ppB02-811,1990.

[%I Y. Gareoich, A~lenge case wmpleteness, J. CCbmpPrt S y d m Sd 42 (1991), pp.346-398.

[3a J. Hartmanis, Generalized KoImogor~~ complexity and the stntcta~e offeasiile computations, in "Pmeeedings, 24th IEEE Symposium on t& Foundations of Camputer Science," pp.439-445, 1983.

[39] J- HartmanLq N. J-, and V. Sdsm, Sparse sets in NP-P: EXPTIUE msw NEXPTIME, Informo- f h 4nd CorrtmI65 (1985), pp.159-181.

[40j J. HSstad, R ImpagliazP,, L A Levin, and M. L d y , Cbnstrnaion of a pseudc+sandom genefator h n any oneway haion, Technics Report, TR-91-068, hkmitional Computer Sdence Institute, BetkeIey, Catifhia,

1991. ,Pd ' j . versions appeared in the pnxxedinp of the 2lst STOC, 1989, pp.12-24 and the 22nd STOC,

1990, pp-39-

[41] M- Hermo and E. Mapordomo, A note on pofynomial-size circuits with Iow resomcebonnded, Moth SystrmJ

Theory, 27 (19941, pp.347-356.

[&I J.E. Kopuoft and J.D. U?man, InfroducCion to Atchnutu Theory, Lunguugu, and Comput4tion, Addison-

WesIey, M a 3 s d m m q 1979.

[43j R ImpagIkiv, A p e r d view of averagecase complexity, in Thmdings, 10th A n n d Structure in b plexity Theory Conferen~e,~ ppJ34-147, 1995.

144 R Impa%iazP, and L A fRoiq No better ways to generate hard NP-instances than picking d o d y at

random, in Tbmd@s, 31th IEEE Symposiam on Foundations of Comprtter Science", pp.812-821, 1990.

[q DS. Johnson, A d o g of complexity dasses, Chapter 2 of Handbook of Tlicorrticof Computa Sdar&, edited by J. van Leewn, Elsevk Science P u b b B.V, 1990.

BIBLIOGRAPHY

PMW. Knijnenbmg, On tillldomkbg decision problems a sm~ejf of the theory of randomized NP, Technkd

Repott RW-CS88-15, Department of Computer Science, Unwersitg of Utrecht, 1988.

K. KO, Some obsmations on the probabilistic algorithms and NP-hard problems, Infbmatim P&g - 14 (19W7 PP-•

K. KO, On some natural complete operations, l7mmtkd Computer Science, 37 (1985), pp.1-30.

K- KO, C o m p w Theory of Reui l h c h n s , BirkhZuser, Boston, 1991.

K. KO and H. Riedman, CompOtationaI complexity of real fmrctioos, Themt id Computer S~~LILC~, 20 (1982),

pp.323-352.

Idf, KO and D. Moore, Completeness, appmxhation and density? SfAM J. Cornpt., 10 (1981), pp.787-796.

sck Kurtz, On the random d e hypothesis, hf- tmd ConLrof 57 (1983), pp.4&47.

L A Levis, Universd seqaenfial search problems, ProbtemJ o f ln f immth l lmmwum . .

9 (1973), pp.26266.

L. Lwb Aoerage case complete probkms, S . J. Compuf 15 (1986), 28S286. A prelimfnary version

appeared under the tide "Problems, compiete m %verage" instance" in "Proceedings, 16th ACM Symposium

on Themy of Computing", p.465,1984.

M. Li and P. M.B. V i Average case complexity under the tmiverd distri'bution eqyah worst-case corn-

p l e 7 Infbmmfim Procwjing Letters, 42 (1992), pp-14H49.

M. Li and P. V i An intduction to Kotmogm cornpkity d i t s uppkatiow, Springer-Vertag, New

York, 1993-

J E Lrrtz, h o s t -here high nonrmifapm compl-, J. of Cumput. Sysfan S d 44 (1992), 220-258.

S R Wq, Sparse complete sets for NP: solution of a conjecture by Bern& and Harmak, J. Canpld

S d 25 (1988), 13&143.

Y. hdansanr, N. Nisan, and P. T i The compat;rtiona a x q h i t y of mdversal hashing, TRuwet Cbmpd

Sd 107 (m), l2l-133-

E Mayordomo, Almost every set in exponential time is P-bi-immune, Tircortt. Cornput. Sci 136 (1994),

487-506.

E. Mayodorno, Contriintions to the study of resource-bouuded measme PED. dissertation, Sistemes In-

f d c s , Universitat Potitecaica de Catalanya, 1994.

A. Me!yer and M. Patefton, Wtth what freqaency are apparently intractable problems difEcuIt ??, Tecknid

Report MIT/LCS/TM-126, MIT, 1979.

A R Meyer and L.J. Stockmeyer, The equidence problem for regalar expressions w i t h qgarhg reqaires

exponential time, in Tmcedkgs, 13th EEE Symposiam on Switching and .4utomata Theory", 125129,

1972

B. Meyer, Constraaive separation of dasses of mdistingaishable ensembles, Trocedhgs, 9th Annual

Confmce on Strachnre in Complexity T h e , pp.198-204,1994.

P.B. Mitterson, The complerxitp of malign d i e s , SIAM J. Cornput, 22 (1993), 147-156.

R. Motwaai a d P. Ra&am~, lbdmid A~~ Cambridge Univezsity Press, New York, 1995.

M. Mundhenk and R Schder? Raadom hgmges for non-dorm complexity chses, J d of Compkdy

7 (1991), -310.

P. Orponen, D A Rtrsso, and tT. Schiining, Optima approDdmations and potynomially Iedable sets, S U M 3-

~mpd 15 (1986), 39948.

P. Orponen and U. Schhing , On the denshy and complexity of po1ylromhI cores for iatrafitable sets, Inform. '

Contml, 70 (t986), pp.5448.

R 0s- and k Wqderson, One-way fanetions are essentS for non-brivia zemimowledge, in "Proceed-

ings? 2nd kd Symposium on Theory of Computing and Systems," Israel, 1993. An extended abstract appeared

T&d I.eport, ICSI-TR-93-073, Mematid Computer Science Institnte at Berkeley, 1993

C. Papadimitdou and S. Zacbw Two remarks on the power of counting, in T m c d b g s , 6th GI Conference

on Theoretkd Computer Science," LectrPe Notes in Computer Science, VoL145, pp.269-275, Springer-Vbrhg,

Berfia, 1983.

BIBLIOGRAPHY

[q Satoh and 0. W- On set size cornpatison problem, TechnicaI Report of IEICE, 1993.

[a] RE. Schap'ae, The emeqhg theory of a- complexity, Technical Report, .MIT/LCS/TM-431, M a s sachusetts fnstitate of Technology, 1990.

w] U. Schiining, Robust algorithms a diffkrent approach to d e s , Cmnputer Sci- 40 (1985),

pp.51-60.

[91] U. Schiining, Complaify cmd dmcbr~, Lectare Notes in Computer Science, V01.211, Springer-VerIag, 1986.

1921 R Schder, h e properties of sets tractable under every polynomial-time computable distriiution, Infdon Pmcmhg kttem, 55 (1995), pp.175bI84.

[93] R SchaIeq Average polynomial time is hard for exponential time under sn-reductions, in 'cProceedings, 15th

Foundations of Softwae TechnoIogg and Theoreti& Computer Sdencqn Lecture Notes in Computer Science,

VoLIO26, pp.240-247,1995.

[94]: R Schder, avth-table dosure aud 'Ihring dmm of amrage polpnomial time have diEerent measure fa EXP, in Troaxcbgs, 11th Anwal Codme on Compdat id CompIexity," pp.19@197,1996.

[95j R Schnlet, A note on anivefta distniations for polynomUtime computable distniution~, in "PKK.nedings, 12th Annuat IEEE Conf-ce on Computational Complexity," 1997.

[96] R W e r and 0. W m Toarards averagecase corn- andpsis of NP optimization probIenrs, in

loth Axutual Conference on Structure in CompIexitp Theory: pp.I4&159,1995.

(9?j R W e t and T. Ya-ai, Structural ;.rtrerage case compIexity, J. Compul. Syskm Sci 52 (1 996), 308-327.

A preliminary version appeared in 12th Foundatians of Software TechnoIogy and Theomtical h p r t t e r %ha%'' hztnre Notes m b p u t e r Science, VoI-65% pp.12&139, Springer-Verlag, BerBerlin, 1992

[q R Schaler and T. Yamak.ami, Sets ampatabIe in polynomial time on awrage, in of the Ist

Computing and -01 Conferen~e,~ Lecture Notes in Computer Science, Vbl-959, pp.400-409,1995,

Springer-Vedag.

[I001 M. S i , A amp- thenretic approach to randomness, in 15th ACM Symposiam on the

Themy of Cornpotink'' pp.33&335,1983-

[I041 L. Valiant, The c o m p l e of computing the permanent, T f r e o d d Computer Science, 5 (1979), 189-201.

[I051 LG. Valiant and V.V. Vazirani, NP is as e q r as detecting unique solutions, Thamtid Cumpute~ Sckm 47

(1=), PP-.

[I061 R V l and L Levin, Random instances of a graph doring pmbIem are hard, in V g s , 20th

ACM Symposium on Theory of Compating", pp.217-222, 1988.

[I01 R Venhesan and S. RaiagopaIan, Atrerage case intractability of Diophantine and matrix problems, in

24th ACM Symposiom on Theory of Computing", pp.632-642, 1992.

[I081 J. Wang, Average case completeness of a word problem for groups, in "Proceedings, 27th ACM Symposium on Theory of Compating", ~~325434 ,1995 .

[110] J. Wang, A m a p a s e W l e NP probIems, manuscript, 1996.

[ill] J. Wang and J. Belangex, On average P vs. average NP, in C m w 27ieory - Cunwaf Rwemrlr, editors K. Ambos-Spies, S. Homer and U. Schihhtg, Cambridge Oniversitp press, 47-61, 1993. A pteliminary version

appeared in 7th IEEE Conference on Structure in C o m p I e Theory," pp.318-326, 1992

[I121 J. Wang and J. Belanger, On the NP-isomorphism probIem with respect to random instances, J. Cornput-

System Sei, 50 (I995), 151-164. A pteliminarp veRion appeared in =gs, 8th Codereace on Structure

in CompIexity Theory,'' 1993, pp.65-74 under the title: rsOmarphisms of .NP compIete prob1ems on madom instances.

[I131 0- Watanabe, Test instance generation far promised NP search problems, in "Proceedings, 9th Anwa Con- f-ce on Structure in Comp1- Theory," pp20&216,1994.

[I161 EU. Wilber, Raa- and the densitp of hard probIems, in "Proceedings, 24th IEEE Symposium on

Foundations of Computer Science", pp.335-342,1983-

[I201 AC. Yao, Theory and appIicatiom of trapdoor frmctians, in Tmad&s, 23rd A d IEEE Symposinm on

Fonnaations of Computer Sciex~ce,~ 1% pp.80-91.

[I211 C. Yap, Some ~mregaences of non-unifarm conditions on aniform dasses, Thcat . Compu+ Sci 26 (1983). pp.287-300.

[I24 S. Zachas, Collapsing probabilistic polynomiaI ltkudk, in "Ebcdhgs , Conference on Cornputatid Complexity Theoayp Santa Barbara, 1983, pp.7Wl.

[I?? S. Zachas, Robabilistic and games, J. Comprt. Syskm Sci 36 (1988), pp.433-451.

List of Notation

M:F ,239 APHF ,239

mgr , 239

APT ,32

q 7 , 239 <wbPP -tt , 185

A=(-(n,n , 79

A=(-(t)?7) ? 79

A=(ATIME~(~, s), 7) , 79

AVCT(ATIME~(~, S) , T ) ,79

A=(-=(T? a, n , 79

A=(-= (t? 4, T ) ? 79 A=(BPP,T)~ ,207

A=(B-(n,T) ? 79

A=(B-(t), 7) ? 79

Aver(C1 nc2,3) ,80

Aver(-GT) ? 80

A=(-Np,7) , 80 Aver(A:, 3) ,217

Aver@spAcE(S), 7) ? 80

A m s p A c E ( 4 , 7 ) , 80

Aver@T=(n?a ? f9

A m T I M E ( t ) , 7 ) , m

LIST OF N O W O N

D=(t) ? 30

E ,30

E-comp , 102

A; 7 33

EH 7 33

q T 33

C; ? 33 ESPACE ,30

I IA(X) ? 275

Q ? 11

EXP-comp ,102

A T ? 33 EXPH ,33

FLAT, 51

Index

........................... symbols t-time computable -42

............. C.f(n)-close .................................. 41 l-dimensional martingale system -45

............................... C-immune 46. 255 .......................... 7 on paverage .55. 65

.................................. Tdomhate 68 ...................... 7-space machine 25

...................... 7-time Thing machine -25

............................. 7-time bounded -29

............................. 7-time machine -29

............................. EXP-computabIe 30

............................... E-cornputabIe -30

....................................... t. rare 110 ............................... s,.c omplete -160 .............................. <,descriptive -167

.................................... <,.hard -160 ............................... P-computable -30

................................... P-printable 36

............... #P-computabIe on paverage -282 ....................................... a-&Id -17

.................................. i-di-5 37 ............................... i-md&iqui& -38

............................ t on pa- 55, 65

tonaverage

with respect to (phIna - ................ 56

with respect to {p+,),, a ................ 56

................................... t- 68

....................... t-space Turing &e 25

.......................... t-space mmptitable -42

................... t-time Tmhg marhine .2S, 29 .............................. &time botmded -29

A

............................... absolute d u e -12

.................................... accept 20, 22

.................. accepting computation tree -23 ....................................... acydic -14

adaptive query .............................. -22 ...................................... adjacent 13

almost

....................................... all 11 .................................... every 11

.............................. everywhere -11

..................................... totd 26

.............................. almost immune 258

....................... aIrnost polynomial time 32

alphabet ..................................... -14 ................................... &ernation 23

..................................... ancestor -14 . .

assoQatrve .................................... 12 ......................... a ~ b p p t t ~ r e d u c i b ~ 185 .......................... average 7dominate 69

........................... average tdominate 69

average polynomial-time ................... many-one redua%w 152

..................... litring redualiritp -158

average poIynomiaIly eqyal ................... 7 l

.................. average poIynomialIy include iT

............... average-polynomia damiaation 69

averagepalynomiaI equidence .............. -69 . . .............................. -n -69

convex ....................................... -16

...................................... covering 12

......................... critical co~gnration -24

........................................ cycle -14

decreasing

strictly ................................... 16 .................................. weakly -16 .................................. degenerative 53

........................................ degree 14

....................................... density 15

.................. den$@ frmction, probabii 50

......................................... depth 14

................................... descendant 14

.................................... di&reIlce -11

................................ disjoint union -15 .................................. distribution -49

............. T-space sampW1e .. ...... -121 ...................... T-time samplab1e -121

............................. T- ttniversal 134

................ EXP-computable .. .... -102

.......................... E-computable -102

.......................... EcomputabIe -102

................ P-comptttabIe .. ......... 102 ........................... P-sampIab1e -121

~&sam~lab1e .......................... -140 ........................ #PamputabIe -131

............. t-space samplab1e ....... ... I21

........................ t-time samplabIe El

.................... average P-sampIabIe I26

a- polynod-time samplabte ...... I26 wnditio rial......................... ...... 50

........... exponential-time compntab1e -102 fIat .................................. 51, 16!3

.................. h m t i i P-sampIabIe 124

hmrtiii1y polyn@time samplabIe .. - 3 4

hear-exponentiaLthne computab1e ...... 102

Iogarithmic-space computable .......... -102

.............................. p u n i d 134

poIynomia1-time computable .. ... .. ..... 102

............. poIynomial-time samplable -121

................................ standard -51

.................. strictly P-computable -103

strictly P+amplable .................... -121

trivial ................................... -49

uoiform ................................. -50

distn'butional problem. ....................... TI

...................................... domain -15

domination condition ....................... -158

dyadic rational number ..... .. ............... -15

E .......................................... edge 13

edgeset ...................................... 13

element ....................................... 11

.................................... empty set 11

.............................. error probability 27

evpnt ......................................... 18

................................... exphonest -16

w o n .............. ........... .......... 18

expected polynomial on paverage ............ -53

expectedvalue ................................ 18

..................... exponential .......... ... -13

exponentidIy bounded ...... .. .............. .. 16

F ....................................... f o d 10

l MAGE EVALUATION TEST TARGET (QA-3)