低血压吃什么好的最快| 脖子长小肉粒是什么原因| 浅表性胃炎吃什么中成药最好| 吸氧有什么好处| 917是什么意思| 湿疹涂什么| s925是什么| 上海话册那是什么意思| 流汗太多对身体有什么危害| 什么是阑尾炎| 副鼻窦炎症是什么意思| 党参不能和什么一起吃| 什么是冠心病| 皮赘用什么药膏去除| 12月1日是什么意思| 虚心接受是什么意思| 2024是什么年生肖| 梦见洗澡是什么意思| 为什么白带是黄色的| 月经不规律是什么原因| 自身免疫性疾病是什么意思| 渠道货是什么意思| 梦见大老鼠是什么意思| 柳字五行属什么| 作梁是什么意思| 秘书是干什么的| 鲱鱼罐头那么臭为什么还有人吃| 球镜是什么意思| 便秘应该挂什么科室| 阳亢是什么意思| 做梦数钱是什么意思啊| 做梦梦到蛆是什么意思| 冰片是什么| 齐人之福什么意思| 宫颈细胞学检查是什么意思| 何以是什么意思| 属马的贵人属相是什么| 根的作用是什么| 渝北区有什么好玩的地方| sy什么意思| 女性分泌物像豆腐渣用什么药| 有趣的灵魂是什么意思| 异常白细胞形态检查是查什么病| 为什么会胃痛| 红艳煞什么意思| 大腿肌肉酸痛是什么病| 一年四季穿棉衣是什么生肖| 克星是什么意思| 尿道痒男吃什么消炎药| 什么是活性叶酸| 婴儿什么时候长牙| 什么是cosplay| 小孩牙龈黑紫色是什么原因| 肺结节吃什么食物散结节最快| 皮肤发黑是什么原因引起的| 异性朋友是什么意思| 浑身乏力什么病的前兆| 爱新觉罗是什么意思| 香草味是什么味道| 内科是什么| 扁桃体发炎有什么症状| 口苦吃什么药最好| 骨头受伤了吃什么恢复的快| lava是什么意思| 肾囊肿有什么症状| 十二生肖分别是什么| 为什么说成也萧何败也萧何| cj什么意思| 男人不举是什么原因造成的| 心脏跳的慢吃什么好| 彪子是什么意思| 眼干是什么原因| 金代表什么数字| 卖剑买牛是什么动物| 体检挂什么科| 目赤什么意思| 蟋蟀吃什么东西| 月经期间喝什么比较好| exo的e为什么不发音| 肺寒吃什么药| 老人嘴唇发紫是什么原因| 肾阴虚吃什么| 扬字五行属什么| 做梦梦到乌龟是什么预兆| 宫颈活检lsil是什么病| 怀孕可以喝什么饮料| 栀子对妇科有什么功效| 大便呈绿色是什么原因| 掌中宝是什么部位| 铅是什么东西| 吃什么能排结石| 凉栀是什么意思| 梦见梅花鹿是什么预兆| 白细胞加号什么意思| 中药为什么那么苦| KTV服务员主要做什么| 口苦吃什么好| 性激素六项查什么| 心有余而力不足什么意思| 女人吃火龙果有什么好处| 早上起床胃疼是什么原因| 梅毒用什么药| 头顶痛吃什么药效果好| 猫头鹰吃什么| 莹字五行属什么| 杜甫世称什么| 晒背有什么好处| 鸟加衣念什么| 呼吸道感染挂什么科| 肚子老是疼是什么原因| 头皮上长疣是什么原因造成的| 育婴师是干什么的| champion什么牌子| 叫床是什么意思| 女人什么年龄性最旺| 吃东西就打嗝是什么原因| 什么烧鸡好吃| 有所作为的意思是什么| 膀胱炎是什么症状表现| 化疗后白细胞低吃什么补得快| 怀孕肚子上长毛是什么原因| 1994属什么| 女性更年期在什么年龄段| 什么是处女膜| 复刻鞋是什么意思| 咳嗽脑袋疼是什么原因| 上眼皮肿是什么原因| 750是什么材质| 英短蓝猫吃什么猫粮好| 男票是什么意思| 耳垂上有痣代表什么| 狂犬疫苗什么时候打有效| 游戏黑洞是什么意思| 什么叫制动| 阿联酋和迪拜什么关系| 虾吃什么| 7月17号是什么星座| 苹果是什么季节的水果| 包皮真菌感染用什么药| 壬子五行属什么| 鹿鞭泡酒有什么功效| 什么是荠菜| 什么是翻墙软件| 8.9是什么星座| 加字五行属什么| 红薯开花预示着什么| 县宣传部长是什么级别| 淋巴结吃什么药| 瘦脱相是什么意思| 情商是什么| 五险一金是指什么| 四川是什么生肖| 西安古代叫什么名字| ca153是什么检查项目| 衡水老白干是什么香型| 汴去掉三点水念什么| 大脚趾发黑是什么原因| 12378是什么电话| 口腔溃疡用什么药| 吃的多拉的少是什么原因| 什么水果清肝火| 为什么摩羯女颜值都高| 女人喝咖啡有什么好处| 胸腔积液是什么意思| 燕窝什么人不适合吃| 什么水是碱性水| offer是什么| 97属什么| 脑梗什么症状| 超导体是什么| 苏打水什么味道| 蓝莓泡酒有什么功效| 孩子头晕挂什么科| 防晒衣的面料是什么| autumn是什么意思| 烟酰胺有什么作用| 饭后胃疼是什么原因| 低钾会有什么症状| jackie是什么意思| 感冒黄鼻涕吃什么药| 构树是什么树| 眼睛双重影什么原因| 耳鸣是什么原因引起的嗡嗡的响| 丹毒是什么病| 静脉注射是什么意思| 百合花语是什么意思| 太多的理由太多的借口是什么歌| 93年属于什么生肖| 桑葚和枸杞泡水喝有什么好处| 过敏性荨麻疹吃什么药| 什么是皮炎| 舌苔白厚吃什么药见效快| 二尖瓣轻度反流是什么意思| 小孩割包皮挂什么科| 胎儿左心室强光斑是什么意思| 什么是肿瘤| 养膘是什么意思| 勾芡用什么粉最好| 速写男装属于什么档次| 什么洗面奶祛痘| 艮是什么意思| 甲亢查什么项目| 狗鼻子为什么是湿的| 鼻涕带血是什么原因引起的| 小孩手指脱皮是什么原因| 濯清涟而不妖的濯是什么意思| 冻顶乌龙茶是什么茶| 蝙蝠侠叫什么| 失眠多梦吃什么药| 一个斤一个页念什么| 人出汗多是什么原因| 虚火是什么意思| 胆囊炎可以吃什么水果| 闲聊是什么意思| 什么是湿气重| 百无一用是什么意思| 喝菊花有什么好处| 阳瘘的最佳治疗方法是什么| 脑供血不足吃什么药效果好| 左手中指痛什么预兆| 第三代身份证什么时候开始办理| 为什么湿气重| 白细胞加号什么意思| 血压高是什么症状| 女属蛇的和什么属相最配| dr是什么检查| 11月5号什么星座| 心肌缺血是什么原因引起的| 非洲说什么语言| 什么手| 拉屎酸臭是什么原因| 梦见被狗咬是什么预兆| 向日葵什么时候种| 法令纹上的痣代表什么| 1995年属什么| 高烧吃什么药| 白居易被称为什么| 名存实亡是什么意思| 开宠物医院需要什么条件| 鸟加衣念什么| 冠心病做什么检查| 为什么会长瘊子| 宫颈口大是什么原因| 手机壳什么材质好| 三个火字读什么| 做梦梦到鬼是什么意思| 软饮是什么意思| 郑和原名叫什么| 齐天大圣是什么意思| 1997年7月1日属什么生肖| 泞字五行属什么| 感冒可以吃什么水果好| 早餐吃什么养胃| 赟读什么| 传媒公司主要做什么| 肚子胀屁多是什么原因| 什么叫数字货币| 冬天有什么水果| 七月一号是什么节| 天秤座男生喜欢什么样的女生| 电脑什么时候发明的| 泄泻病是什么意思| 切糕为什么这么贵| 浅表性胃炎是什么意思| 百度Jump to content

载192人轮渡在韩国西南海岸触礁 6人受轻伤韩国轮渡触礁

From Wikipedia, the free encyclopedia
(Redirected from Tractable problem)
百度 为减肥,网贷的7万元没用,却借钱还利息毕竟阿欣是个90后女孩,不甘心一直胖下去。

In theoretical computer science and mathematics, computational complexity theory focuses on classifying computational problems according to their resource usage, and explores the relationships between these classifications. A computational problem is a task solved by a computer. A computation problem is solvable by mechanical application of mathematical steps, such as an algorithm.

A problem is regarded as inherently difficult if its solution requires significant resources, whatever the algorithm used. The theory formalizes this intuition, by introducing mathematical models of computation to study these problems and quantifying their computational complexity, i.e., the amount of resources needed to solve them, such as time and storage. Other measures of complexity are also used, such as the amount of communication (used in communication complexity), the number of gates in a circuit (used in circuit complexity) and the number of processors (used in parallel computing). One of the roles of computational complexity theory is to determine the practical limits on what computers can and cannot do. The P versus NP problem, one of the seven Millennium Prize Problems,[1] is part of the field of computational complexity.

Closely related fields in theoretical computer science are analysis of algorithms and computability theory. A key distinction between analysis of algorithms and computational complexity theory is that the former is devoted to analyzing the amount of resources needed by a particular algorithm to solve a problem, whereas the latter asks a more general question about all possible algorithms that could be used to solve the same problem. More precisely, computational complexity theory tries to classify problems that can or cannot be solved with appropriately restricted resources. In turn, imposing restrictions on the available resources is what distinguishes computational complexity from computability theory: the latter theory asks what kinds of problems can, in principle, be solved algorithmically.

Computational problems

[edit]
A traveling salesman tour through 14 German cities

Problem instances

[edit]

A computational problem can be viewed as an infinite collection of instances together with a set (possibly empty) of solutions for every instance. The input string for a computational problem is referred to as a problem instance, and should not be confused with the problem itself. In computational complexity theory, a problem refers to the abstract question to be solved. In contrast, an instance of this problem is a rather concrete utterance, which can serve as the input for a decision problem. For example, consider the problem of primality testing. The instance is a number (e.g., 15) and the solution is "yes" if the number is prime and "no" otherwise (in this case, 15 is not prime and the answer is "no"). Stated another way, the instance is a particular input to the problem, and the solution is the output corresponding to the given input.

To further highlight the difference between a problem and an instance, consider the following instance of the decision version of the travelling salesman problem: Is there a route of at most 2000 kilometres passing through all of Germany's 14 largest cities? The quantitative answer to this particular problem instance is of little use for solving other instances of the problem, such as asking for a round trip through all sites in Milan whose total length is at most 10 km. For this reason, complexity theory addresses computational problems and not particular problem instances.

Representing problem instances

[edit]

When considering computational problems, a problem instance is a string over an alphabet. Usually, the alphabet is taken to be the binary alphabet (i.e., the set {0,1}), and thus the strings are bitstrings. As in a real-world computer, mathematical objects other than bitstrings must be suitably encoded. For example, integers can be represented in binary notation, and graphs can be encoded directly via their adjacency matrices, or by encoding their adjacency lists in binary.

Even though some proofs of complexity-theoretic theorems regularly assume some concrete choice of input encoding, one tries to keep the discussion abstract enough to be independent of the choice of encoding. This can be achieved by ensuring that different representations can be transformed into each other efficiently.

Decision problems as formal languages

[edit]
A decision problem has only two possible outputs, yes or no (or alternately 1 or 0) on any input.

Decision problems are one of the central objects of study in computational complexity theory. A decision problem is a type of computational problem where the answer is either yes or no (alternatively, 1 or 0). A decision problem can be viewed as a formal language, where the members of the language are instances whose output is yes, and the non-members are those instances whose output is no. The objective is to decide, with the aid of an algorithm, whether a given input string is a member of the formal language under consideration. If the algorithm deciding this problem returns the answer yes, the algorithm is said to accept the input string, otherwise it is said to reject the input.

An example of a decision problem is the following. The input is an arbitrary graph. The problem consists in deciding whether the given graph is connected or not. The formal language associated with this decision problem is then the set of all connected graphs — to obtain a precise definition of this language, one has to decide how graphs are encoded as binary strings.

Function problems

[edit]

A function problem is a computational problem where a single output (of a total function) is expected for every input, but the output is more complex than that of a decision problem—that is, the output is not just yes or no. Notable examples include the traveling salesman problem and the integer factorization problem.

It is tempting to think that the notion of function problems is much richer than the notion of decision problems. However, this is not really the case, since function problems can be recast as decision problems. For example, the multiplication of two integers can be expressed as the set of triples such that the relation holds. Deciding whether a given triple is a member of this set corresponds to solving the problem of multiplying two numbers.

Measuring the size of an instance

[edit]

To measure the difficulty of solving a computational problem, one may wish to see how much time the best algorithm requires to solve the problem. However, the running time may, in general, depend on the instance. In particular, larger instances will require more time to solve. Thus the time required to solve a problem (or the space required, or any measure of complexity) is calculated as a function of the size of the instance. The input size is typically measured in bits. Complexity theory studies how algorithms scale as input size increases. For instance, in the problem of finding whether a graph is connected, how much more time does it take to solve a problem for a graph with vertices compared to the time taken for a graph with vertices?

If the input size is , the time taken can be expressed as a function of . Since the time taken on different inputs of the same size can be different, the worst-case time complexity is defined to be the maximum time taken over all inputs of size . If is a polynomial in , then the algorithm is said to be a polynomial time algorithm. Cobham's thesis argues that a problem can be solved with a feasible amount of resources if it admits a polynomial-time algorithm.

Machine models and complexity measures

[edit]

Turing machine

[edit]
An illustration of a Turing machine

A Turing machine is a mathematical model of a general computing machine. It is a theoretical device that manipulates symbols contained on a strip of tape. Turing machines are not intended as a practical computing technology, but rather as a general model of a computing machine—anything from an advanced supercomputer to a mathematician with a pencil and paper. It is believed that if a problem can be solved by an algorithm, there exists a Turing machine that solves the problem. Indeed, this is the statement of the Church–Turing thesis. Furthermore, it is known that everything that can be computed on other models of computation known to us today, such as a RAM machine, Conway's Game of Life, cellular automata, lambda calculus or any programming language can be computed on a Turing machine. Since Turing machines are easy to analyze mathematically, and are believed to be as powerful as any other model of computation, the Turing machine is the most commonly used model in complexity theory.

Many types of Turing machines are used to define complexity classes, such as deterministic Turing machines, probabilistic Turing machines, non-deterministic Turing machines, quantum Turing machines, symmetric Turing machines and alternating Turing machines. They are all equally powerful in principle, but when resources (such as time or space) are bounded, some of these may be more powerful than others.

A deterministic Turing machine is the most basic Turing machine, which uses a fixed set of rules to determine its future actions. A probabilistic Turing machine is a deterministic Turing machine with an extra supply of random bits. The ability to make probabilistic decisions often helps algorithms solve problems more efficiently. Algorithms that use random bits are called randomized algorithms. A non-deterministic Turing machine is a deterministic Turing machine with an added feature of non-determinism, which allows a Turing machine to have multiple possible future actions from a given state. One way to view non-determinism is that the Turing machine branches into many possible computational paths at each step, and if it solves the problem in any of these branches, it is said to have solved the problem. Clearly, this model is not meant to be a physically realizable model, it is just a theoretically interesting abstract machine that gives rise to particularly interesting complexity classes. For examples, see non-deterministic algorithm.

Other machine models

[edit]

Many machine models different from the standard multi-tape Turing machines have been proposed in the literature, for example random-access machines. Perhaps surprisingly, each of these models can be converted to another without providing any extra computational power. The time and memory consumption of these alternate models may vary.[2] What all these models have in common is that the machines operate deterministically.

However, some computational problems are easier to analyze in terms of more unusual resources. For example, a non-deterministic Turing machine is a computational model that is allowed to branch out to check many different possibilities at once. The non-deterministic Turing machine has very little to do with how we physically want to compute algorithms, but its branching exactly captures many of the mathematical models we want to analyze, so that non-deterministic time is a very important resource in analyzing computational problems.

Complexity measures

[edit]

For a precise definition of what it means to solve a problem using a given amount of time and space, a computational model such as the deterministic Turing machine is used. The time required by a deterministic Turing machine on input is the total number of state transitions, or steps, the machine makes before it halts and outputs the answer ("yes" or "no"). A Turing machine is said to operate within time if the time required by on each input of length is at most . A decision problem can be solved in time if there exists a Turing machine operating in time that solves the problem. Since complexity theory is interested in classifying problems based on their difficulty, one defines sets of problems based on some criteria. For instance, the set of problems solvable within time on a deterministic Turing machine is then denoted by DTIME().

Analogous definitions can be made for space requirements. Although time and space are the most well-known complexity resources, any complexity measure can be viewed as a computational resource. Complexity measures are very generally defined by the Blum complexity axioms. Other complexity measures used in complexity theory include communication complexity, circuit complexity, and decision tree complexity.

The complexity of an algorithm is often expressed using big O notation.

Best, worst and average case complexity

[edit]
Visualization of the quicksort algorithm, which has average case performance

The best, worst and average case complexity refer to three different ways of measuring the time complexity (or any other complexity measure) of different inputs of the same size. Since some inputs of size may be faster to solve than others, we define the following complexities:

  1. Best-case complexity: This is the complexity of solving the problem for the best input of size .
  2. Average-case complexity: This is the complexity of solving the problem on an average. This complexity is only defined with respect to a probability distribution over the inputs. For instance, if all inputs of the same size are assumed to be equally likely to appear, the average case complexity can be defined with respect to the uniform distribution over all inputs of size .
  3. Amortized analysis: Amortized analysis considers both the costly and less costly operations together over the whole series of operations of the algorithm.
  4. Worst-case complexity: This is the complexity of solving the problem for the worst input of size .

The order from cheap to costly is: Best, average (of discrete uniform distribution), amortized, worst.

For example, the deterministic sorting algorithm quicksort addresses the problem of sorting a list of integers. The worst-case is when the pivot is always the largest or smallest value in the list (so the list is never divided). In this case, the algorithm takes time O(). If we assume that all possible permutations of the input list are equally likely, the average time taken for sorting is . The best case occurs when each pivoting divides the list in half, also needing time.

Upper and lower bounds on the complexity of problems

[edit]

To classify the computation time (or similar resources, such as space consumption), it is helpful to demonstrate upper and lower bounds on the maximum amount of time required by the most efficient algorithm to solve a given problem. The complexity of an algorithm is usually taken to be its worst-case complexity unless specified otherwise. Analyzing a particular algorithm falls under the field of analysis of algorithms. To show an upper bound on the time complexity of a problem, one needs to show only that there is a particular algorithm with running time at most . However, proving lower bounds is much more difficult, since lower bounds make a statement about all possible algorithms that solve a given problem. The phrase "all possible algorithms" includes not just the algorithms known today, but any algorithm that might be discovered in the future. To show a lower bound of for a problem requires showing that no algorithm can have time complexity lower than .

Upper and lower bounds are usually stated using the big O notation, which hides constant factors and smaller terms. This makes the bounds independent of the specific details of the computational model used. For instance, if , in big O notation one would write .


Complexity classes

[edit]

Defining complexity classes

[edit]

A complexity class is a set of problems of related complexity. Simpler complexity classes are defined by the following factors:

  • The type of computational problem: The most commonly used problems are decision problems. However, complexity classes can be defined based on function problems, counting problems, optimization problems, promise problems, etc.
  • The model of computation: The most common model of computation is the deterministic Turing machine, but many complexity classes are based on non-deterministic Turing machines, Boolean circuits, quantum Turing machines, monotone circuits, etc.
  • The resource (or resources) that is being bounded and the bound: These two properties are usually stated together, such as "polynomial time", "logarithmic space", "constant depth", etc.

Some complexity classes have complicated definitions that do not fit into this framework. Thus, a typical complexity class has a definition like the following:

The set of decision problems solvable by a deterministic Turing machine within time . (This complexity class is known as DTIME().)

But bounding the computation time above by some concrete function often yields complexity classes that depend on the chosen machine model. For instance, the language can be solved in linear time on a multi-tape Turing machine, but necessarily requires quadratic time in the model of single-tape Turing machines. If we allow polynomial variations in running time, Cobham-Edmonds thesis states that "the time complexities in any two reasonable and general models of computation are polynomially related" (Goldreich 2008, Chapter 1.2). This forms the basis for the complexity class P, which is the set of decision problems solvable by a deterministic Turing machine within polynomial time. The corresponding set of function problems is FP.

Important complexity classes

[edit]
A representation of the relation among complexity classes; L would be another step "inside" NL

Many important complexity classes can be defined by bounding the time or space used by the algorithm. Some important complexity classes of decision problems defined in this manner are the following:

Resource Determinism Complexity class Resource constraint
Space Non-Deterministic NSPACE()
NL
NPSPACE
NEXPSPACE
Deterministic DSPACE()
L
PSPACE
EXPSPACE
Time Non-Deterministic NTIME()
NP
NEXPTIME
Deterministic DTIME()
P
EXPTIME

Logarithmic-space classes do not account for the space required to represent the problem.

It turns out that PSPACE = NPSPACE and EXPSPACE = NEXPSPACE by Savitch's theorem.

Other important complexity classes include BPP, ZPP and RP, which are defined using probabilistic Turing machines; AC and NC, which are defined using Boolean circuits; and BQP and QMA, which are defined using quantum Turing machines. #P is an important complexity class of counting problems (not decision problems). Classes like IP and AM are defined using Interactive proof systems. ALL is the class of all decision problems.

Hierarchy theorems

[edit]

For the complexity classes defined in this way, it is desirable to prove that relaxing the requirements on (say) computation time indeed defines a bigger set of problems. In particular, although DTIME() is contained in DTIME(), it would be interesting to know if the inclusion is strict. For time and space requirements, the answer to such questions is given by the time and space hierarchy theorems respectively. They are called hierarchy theorems because they induce a proper hierarchy on the classes defined by constraining the respective resources. Thus there are pairs of complexity classes such that one is properly included in the other. Having deduced such proper set inclusions, we can proceed to make quantitative statements about how much more additional time or space is needed in order to increase the number of problems that can be solved.

More precisely, the time hierarchy theorem states that .

The space hierarchy theorem states that .

The time and space hierarchy theorems form the basis for most separation results of complexity classes. For instance, the time hierarchy theorem tells us that P is strictly contained in EXPTIME, and the space hierarchy theorem tells us that L is strictly contained in PSPACE.

Reduction

[edit]

Many complexity classes are defined using the concept of a reduction. A reduction is a transformation of one problem into another problem. It captures the informal notion of a problem being at most as difficult as another problem. For instance, if a problem can be solved using an algorithm for , is no more difficult than , and we say that reduces to . There are many different types of reductions, based on the method of reduction, such as Cook reductions, Karp reductions and Levin reductions, and the bound on the complexity of reductions, such as polynomial-time reductions or log-space reductions.

The most commonly used reduction is a polynomial-time reduction. This means that the reduction process takes polynomial time. For example, the problem of squaring an integer can be reduced to the problem of multiplying two integers. This means an algorithm for multiplying two integers can be used to square an integer. Indeed, this can be done by giving the same input to both inputs of the multiplication algorithm. Thus we see that squaring is not more difficult than multiplication, since squaring can be reduced to multiplication.

This motivates the concept of a problem being hard for a complexity class. A problem is hard for a class of problems if every problem in can be reduced to . Thus no problem in is harder than , since an algorithm for allows us to solve any problem in . The notion of hard problems depends on the type of reduction being used. For complexity classes larger than P, polynomial-time reductions are commonly used. In particular, the set of problems that are hard for NP is the set of NP-hard problems.

If a problem is in and hard for , then is said to be complete for . This means that is the hardest problem in . (Since many problems could be equally hard, one might say that is one of the hardest problems in .) Thus the class of NP-complete problems contains the most difficult problems in NP, in the sense that they are the ones most likely not to be in P. Because the problem P = NP is not solved, being able to reduce a known NP-complete problem, , to another problem, , would indicate that there is no known polynomial-time solution for . This is because a polynomial-time solution to would yield a polynomial-time solution to . Similarly, because all NP problems can be reduced to the set, finding an NP-complete problem that can be solved in polynomial time would mean that P = NP.[3]

Important open problems

[edit]
Diagram of complexity classes provided that P ≠ NP. The existence of problems in NP outside both P and NP-complete in this case was established by Ladner.[4]

P versus NP problem

[edit]

The complexity class P is often seen as a mathematical abstraction modeling those computational tasks that admit an efficient algorithm. This hypothesis is called the Cobham–Edmonds thesis. The complexity class NP, on the other hand, contains many problems that people would like to solve efficiently, but for which no efficient algorithm is known, such as the Boolean satisfiability problem, the Hamiltonian path problem and the vertex cover problem. Since deterministic Turing machines are special non-deterministic Turing machines, it is easily observed that each problem in P is also member of the class NP.

The question of whether P equals NP is one of the most important open questions in theoretical computer science because of the wide implications of a solution.[3] If the answer is yes, many important problems can be shown to have more efficient solutions. These include various types of integer programming problems in operations research, many problems in logistics, protein structure prediction in biology,[5] and the ability to find formal proofs of pure mathematics theorems.[6] The P versus NP problem is one of the Millennium Prize Problems proposed by the Clay Mathematics Institute. There is a US$1,000,000 prize for resolving the problem.[7]

Problems in NP not known to be in P or NP-complete

[edit]

It was shown by Ladner that if then there exist problems in that are neither in nor -complete.[4] Such problems are called NP-intermediate problems. The graph isomorphism problem, the discrete logarithm problem and the integer factorization problem are examples of problems believed to be NP-intermediate. They are some of the very few NP problems not known to be in or to be -complete.

The graph isomorphism problem is the computational problem of determining whether two finite graphs are isomorphic. An important unsolved problem in complexity theory is whether the graph isomorphism problem is in , -complete, or NP-intermediate. The answer is not known, but it is believed that the problem is at least not NP-complete.[8] If graph isomorphism is NP-complete, the polynomial time hierarchy collapses to its second level.[9] Since it is widely believed that the polynomial hierarchy does not collapse to any finite level, it is believed that graph isomorphism is not NP-complete. The best algorithm for this problem, due to László Babai and Eugene Luks has run time for graphs with vertices, although some recent work by Babai offers some potentially new perspectives on this.[10]

The integer factorization problem is the computational problem of determining the prime factorization of a given integer. Phrased as a decision problem, it is the problem of deciding whether the input has a prime factor less than . No efficient integer factorization algorithm is known, and this fact forms the basis of several modern cryptographic systems, such as the RSA algorithm. The integer factorization problem is in and in (and even in UP and co-UP[11]). If the problem is -complete, the polynomial time hierarchy will collapse to its first level (i.e., will equal ). The best known algorithm for integer factorization is the general number field sieve, which takes time [12] to factor an odd integer . However, the best known quantum algorithm for this problem, Shor's algorithm, does run in polynomial time. Unfortunately, this fact doesn't say much about where the problem lies with respect to non-quantum complexity classes.

Separations between other complexity classes

[edit]

Many known complexity classes are suspected to be unequal, but this has not been proved. For instance , but it is possible that . If is not equal to , then is not equal to either. Since there are many known complexity classes between and , such as , , , , , , etc., it is possible that all these complexity classes collapse to one class. Proving that any of these classes are unequal would be a major breakthrough in complexity theory.

Along the same lines, is the class containing the complement problems (i.e. problems with the yes/no answers reversed) of problems. It is believed[13] that is not equal to ; however, it has not yet been proven. It is clear that if these two complexity classes are not equal then is not equal to , since . Thus if we would have whence .

Similarly, it is not known if (the set of all problems that can be solved in logarithmic space) is strictly contained in or equal to . Again, there are many complexity classes between the two, such as and , and it is not known if they are distinct or equal classes.

It is suspected that and are equal. However, it is currently open if .

Intractability

[edit]

A problem that can theoretically be solved, but requires impractical and infinite resources (e.g., time) to do so, is known as an intractable problem.[14] Conversely, a problem that can be solved in practice is called a tractable problem, literally "a problem that can be handled". The term infeasible (literally "cannot be done") is sometimes used interchangeably with intractable,[15] though this risks confusion with a feasible solution in mathematical optimization.[16]

Tractable problems are frequently identified with problems that have polynomial-time solutions (, ); this is known as the Cobham–Edmonds thesis. Problems that are known to be intractable in this sense include those that are EXPTIME-hard. If is not the same as , then NP-hard problems are also intractable in this sense.

However, this identification is inexact: a polynomial-time solution with large degree or large leading coefficient grows quickly, and may be impractical for practical size problems; conversely, an exponential-time solution that grows slowly may be practical on realistic input, or a solution that takes a long time in the worst case may take a short time in most cases or the average case, and thus still be practical. Saying that a problem is not in does not imply that all large cases of the problem are hard or even that most of them are. For example, the decision problem in Presburger arithmetic has been shown not to be in , yet algorithms have been written that solve the problem in reasonable times in most cases. Similarly, algorithms can solve the NP-complete knapsack problem over a wide range of sizes in less than quadratic time and SAT solvers routinely handle large instances of the NP-complete Boolean satisfiability problem.

To see why exponential-time algorithms are generally unusable in practice, consider a program that makes operations before halting. For small , say 100, and assuming for the sake of example that the computer does operations each second, the program would run for about years, which is the same order of magnitude as the age of the universe. Even with a much faster computer, the program would only be useful for very small instances and in that sense the intractability of a problem is somewhat independent of technological progress. However, an exponential-time algorithm that takes operations is practical until gets relatively large.

Similarly, a polynomial time algorithm is not always practical. If its running time is, say, , it is unreasonable to consider it efficient and it is still useless except on small instances. Indeed, in practice even or algorithms are often impractical on realistic sizes of problems.

Continuous complexity theory

[edit]

Continuous complexity theory can refer to complexity theory of problems that involve continuous functions that are approximated by discretizations, as studied in numerical analysis. One approach to complexity theory of numerical analysis[17] is information based complexity.

Continuous complexity theory can also refer to complexity theory of the use of analog computation, which uses continuous dynamical systems and differential equations.[18] Control theory can be considered a form of computation and differential equations are used in the modelling of continuous-time and hybrid discrete-continuous-time systems.[19]

History

[edit]

An early example of algorithm complexity analysis is the running time analysis of the Euclidean algorithm done by Gabriel Lamé in 1844.

Before the actual research explicitly devoted to the complexity of algorithmic problems started off, numerous foundations were laid out by various researchers. Most influential among these was the definition of Turing machines by Alan Turing in 1936, which turned out to be a very robust and flexible simplification of a computer.

The beginning of systematic studies in computational complexity is attributed to the seminal 1965 paper "On the Computational Complexity of Algorithms" by Juris Hartmanis and Richard E. Stearns, which laid out the definitions of time complexity and space complexity, and proved the hierarchy theorems.[20] In addition, in 1965 Edmonds suggested to consider a "good" algorithm to be one with running time bounded by a polynomial of the input size.[21]

Earlier papers studying problems solvable by Turing machines with specific bounded resources include[20] John Myhill's definition of linear bounded automata (Myhill 1960), Raymond Smullyan's study of rudimentary sets (1961), as well as Hisao Yamada's paper[22] on real-time computations (1962). Somewhat earlier, Boris Trakhtenbrot (1956), a pioneer in the field from the USSR, studied another specific complexity measure.[23] As he remembers:

However, [my] initial interest [in automata theory] was increasingly set aside in favor of computational complexity, an exciting fusion of combinatorial methods, inherited from switching theory, with the conceptual arsenal of the theory of algorithms. These ideas had occurred to me earlier in 1955 when I coined the term "signalizing function", which is nowadays commonly known as "complexity measure".[24]

In 1967, Manuel Blum formulated a set of axioms (now known as Blum axioms) specifying desirable properties of complexity measures on the set of computable functions and proved an important result, the so-called speed-up theorem. The field began to flourish in 1971 when Stephen Cook and Leonid Levin proved the existence of practically relevant problems that are NP-complete. In 1972, Richard Karp took this idea a leap forward with his landmark paper, "Reducibility Among Combinatorial Problems", in which he showed that 21 diverse combinatorial and graph theoretical problems, each infamous for its computational intractability, are NP-complete.[25]

See also

[edit]

Works on complexity

[edit]
  • Wuppuluri, Shyam; Doria, Francisco A., eds. (2020), Unravelling Complexity: The Life and Work of Gregory Chaitin, World Scientific, doi:10.1142/11270, ISBN 978-981-12-0006-9, S2CID 198790362

References

[edit]

Citations

[edit]
  1. ^ "P vs NP Problem | Clay Mathematics Institute". www.claymath.org. Archived from the original on July 6, 2018. Retrieved July 6, 2018.
  2. ^ See Arora & Barak 2009, Chapter 1: The computational model and why it doesn't matter
  3. ^ a b See Sipser 2006, Chapter 7: Time complexity
  4. ^ a b Ladner, Richard E. (1975), "On the structure of polynomial time reducibility", Journal of the ACM, 22 (1): 151–171, doi:10.1145/321864.321877, S2CID 14352974.
  5. ^ Berger, Bonnie A.; Leighton, T (1998), "Protein folding in the hydrophobic-hydrophilic (HP) model is NP-complete", Journal of Computational Biology, 5 (1): 27–40, CiteSeerX 10.1.1.139.5547, doi:10.1089/cmb.1998.5.27, PMID 9541869.
  6. ^ Cook, Stephen (April 2000), The P versus NP Problem (PDF), Clay Mathematics Institute, archived from the original (PDF) on December 12, 2010, retrieved October 18, 2006.
  7. ^ Jaffe, Arthur M. (2006), "The Millennium Grand Challenge in Mathematics" (PDF), Notices of the AMS, 53 (6), archived (PDF) from the original on June 12, 2006, retrieved October 18, 2006.
  8. ^ Arvind, Vikraman; Kurur, Piyush P. (2006), "Graph isomorphism is in SPP", Information and Computation, 204 (5): 835–852, doi:10.1016/j.ic.2006.02.002.
  9. ^ Sch?ning, Uwe (1988), "Graph Isomorphism is in the Low Hierarchy", Journal of Computer and System Sciences, 37 (3): 312–323, doi:10.1016/0022-0000(88)90010-4
  10. ^ Babai, László (2016). "Graph Isomorphism in Quasipolynomial Time". arXiv:1512.03547 [cs.DS].
  11. ^ Fortnow, Lance (September 13, 2002). "Computational Complexity Blog: Factoring". weblog.fortnow.com.
  12. ^ Wolfram MathWorld: Number Field Sieve
  13. ^ Boaz Barak's course on Computational Complexity Lecture 2
  14. ^ Hopcroft, J.E., Motwani, R. and Ullman, J.D. (2007) Introduction to Automata Theory, Languages, and Computation, Addison Wesley, Boston/San Francisco/New York (page 368)
  15. ^ Meurant, Gerard (2014). Algorithms and Complexity. Elsevier. p. p. 4. ISBN 978-0-08093391-7.
  16. ^ Zobel, Justin (2015). Writing for Computer Science. Springer. p. 132. ISBN 978-1-44716639-9.
  17. ^ Smale, Steve (1997). "Complexity Theory and Numerical Analysis". Acta Numerica. 6. Cambridge Univ Press: 523–551. Bibcode:1997AcNum...6..523S. CiteSeerX 10.1.1.33.4678. doi:10.1017/s0962492900002774. S2CID 5949193.
  18. ^ Babai, László; Campagnolo, Manuel (2009). "A Survey on Continuous Time Computations". arXiv:0907.3117 [cs.CC].
  19. ^ Tomlin, Claire J.; Mitchell, Ian; Bayen, Alexandre M.; Oishi, Meeko (July 2003). "Computational Techniques for the Verification of Hybrid Systems". Proceedings of the IEEE. 91 (7): 986–1001. CiteSeerX 10.1.1.70.4296. doi:10.1109/jproc.2003.814621.
  20. ^ a b Fortnow & Homer (2003)
  21. ^ Richard M. Karp, "Combinatorics, Complexity, and Randomness", 1985 Turing Award Lecture
  22. ^ Yamada, H. (1962). "Real-Time Computation and Recursive Functions Not Real-Time Computable". IEEE Transactions on Electronic Computers. EC-11 (6): 753–760. doi:10.1109/TEC.1962.5219459.
  23. ^ Trakhtenbrot, B.A.: Signalizing functions and tabular operators. Uchionnye Zapiski Penzenskogo Pedinstituta (Transactions of the Penza Pedagogoical Institute) 4, 75–87 (1956) (in Russian)
  24. ^ Boris Trakhtenbrot, "From Logic to Theoretical Computer Science – An Update". In: Pillars of Computer Science, LNCS 4800, Springer 2008.
  25. ^ Richard M. Karp (1972), "Reducibility Among Combinatorial Problems" (PDF), in R. E. Miller; J. W. Thatcher (eds.), Complexity of Computer Computations, New York: Plenum, pp. 85–103, archived from the original (PDF) on June 29, 2011, retrieved September 28, 2009{{citation}}: CS1 maint: publisher location (link)

Textbooks

[edit]

Surveys

[edit]
[edit]
惊讶的什么 肚子咕咕叫是什么原因 水猴子是什么动物 更年期是什么时候 bp在医学上是什么意思
上午九点到十一点是什么时辰 财神位放什么最旺财 人格魅力什么意思 食管炎是什么原因引起的 什么是人设
醋坛子是什么意思 蚊子为什么会咬人 世界上最大的动物是什么 五月二十九是什么日子 胃切除有什么影响
玉露茶属于什么茶 黑苦荞茶有什么功效 纯碱是什么 pr是什么职位 粗糙的什么
杏色配什么颜色好看kuyehao.com 626什么星座hcv8jop4ns2r.cn 喝茶叶水有什么好处和坏处hcv8jop4ns9r.cn 忌廉是什么东西hcv7jop4ns6r.cn 疖肿是什么样子的图片hcv8jop7ns5r.cn
眼角痒是什么原因hcv9jop3ns8r.cn 尿道感染用什么消炎药hcv9jop1ns0r.cn 藏青和藏蓝有什么区别hcv9jop4ns9r.cn 0到3个月的婴儿惊吓吃什么药hcv8jop9ns1r.cn 负数是什么意思hcv8jop6ns1r.cn
儿童便秘吃什么最快排便zhiyanzhang.com 6.19什么星座hcv9jop3ns9r.cn 333是什么意思hcv8jop0ns7r.cn bp是什么单位youbangsi.com 声嘶力竭是什么意思hcv9jop6ns8r.cn
来例假可以吃什么水果hcv9jop6ns1r.cn 吃避孕药为什么要吃维生素chcv9jop0ns7r.cn 一什么教室onlinewuye.com 隔三差五是什么意思hcv8jop3ns6r.cn rl是什么单位hcv8jop6ns7r.cn
百度