梦见撒尿是什么意思| 私密是什么意思| 京东白条什么时候还款| 白细胞低吃什么药| 植物园里有什么| 总胆红素偏高吃什么药| 女生什么时候绝经| 水瓶座与什么星座最配| triangle是什么意思| 一见钟情是什么感觉| 石家庄有什么特产| 为什么青蛙跳的比树高| 学信网上的报告编号是什么| 毛峰茶属于什么茶| 藩王是什么意思| 前列腺炎忌口什么食物| 感冒可以吃什么| 成功是什么| tory burch什么牌子| 森达属于什么档次的鞋| 补气血吃什么水果| 六字真言是什么| 绰号是什么意思| 基酒是什么意思| 维生素B6有什么功效| 淋巴挂什么科| 卿卿什么意思| 什么样的礼物| 脉细是什么意思| 舌苔白吃什么药效果好| 小肚子发胀是什么原因女性| 气短挂什么科| 疾病是什么意思| 女性感染hpv有什么症状| 发烧想吐是什么原因| 布鲁斯是什么| 牙龈为什么会肿| 溃疡用什么药| 梦见建房子是什么预兆| 兔唇是什么原因造成的| 西昌火把节是什么时候| 吹空调嗓子疼吃什么药| 身败名裂是什么意思| 一般什么人容易得甲亢| 珍珠鸟吃什么| 脑血管堵塞吃什么药好| 梦见不干净的东西代表什么| 低血糖看什么科室| 耳朵痛什么原因| 元神是什么意思| 教皇是什么意思| 不能吃辛辣是指什么| 家里什么东西止血最快| 龙眼和桂圆有什么区别| 5月6日什么星座| 发烧一直不退是什么原因| 嘴角上火是什么原因| 日出扶桑是什么意思| vogue是什么意思| dsd是什么意思| 金刚菩提是什么植物的种子| 吃什么疏通血管| 铃字五行属什么| 心电图窦性心律是什么意思| 故友是什么意思| 虞是什么意思| 难为你了是什么意思| gold是什么牌子| 三七长什么样| 胸椎退行性变什么意思| 巨无霸是什么意思| 颈椎引起的头晕是什么症状| 八卦什么意思| 沙僧的武器是什么| 口干舌燥口苦吃什么药| 花开花落不见你回头是什么歌| 鬼是什么意思| 什么是狂躁症| 刺身是什么鱼| 中华文化的精髓是什么| 双肺结节是什么病| 黄历修造是什么意思| 企鹅吃什么| 淼读什么字| 怎么知道自己五行缺什么| 小孩出汗多是什么原因| 什么是汛期| 熊猫为什么叫熊猫| 刘备的儿子叫什么| 鼻窦炎用什么药效果最好| syp是什么意思| 办慢性病有什么好处| 三铵复合肥是什么| 花中皇后指的是什么花| 白茶和绿茶有什么区别| 得了破伤风是什么症状| 鸡的五行属什么| 桃花灼灼什么意思| 地球是什么星| 憩是什么意思| 教师编制是什么意思| 诺迪康胶囊治什么病| 小葱拌豆腐的下一句是什么| 生吃大葱有什么好处和坏处| 2023年属兔的是什么命| 肝斑一般在脸上的什么地方| 温柔的动物是什么生肖| 图谱是什么意思| 镜花缘是什么意思| 中午吃什么| 非转基因是什么意思| 为什么不敢挖雍正陵墓| 乳头痒用什么药| 功能性消化不良是什么意思| 布克兄弟什么档次| 氨咖黄敏胶囊是什么药| 三七治什么病最好| 毛躁是什么意思| 早上口干苦是什么原因| 月关是什么意思| 低血糖吃什么最快缓解| 刘备代表什么生肖| 甲抗是什么原因引起的| 慢性肠炎吃什么药最好| 幼儿反复发烧是什么原因| 枕头太低了有什么危害| 头晕是什么病的征兆| 胡萝卜不能和什么一起吃| 吃什么药降尿酸快| 性无能是什么意思| 稀疏是什么意思| 肾萎缩是什么原因引起的| 十月十日什么星座| 6合是什么生肖| 带状疱疹是什么原因引起| 血糖高吃什么饭| 唾液臭是什么原因| 养肝护肝吃什么最好| 盘尼西林是什么药| 叩首是什么意思| 什么是津液| 糖尿病能吃什么零食| 噫气是什么意思| 得了艾滋病有什么症状| 二脚趾比大脚趾长代表什么| 2016属什么生肖| 96年出生的属什么| 生理期不能吃什么水果| 一个虫一个尧念什么| 乙醚是什么| 珈字五行属什么| 儿童乘坐飞机需要什么证件| 经常胃胀气是什么原因引起的| 尿粒细胞酯酶阳性什么意思| 嘴唇神经跳动是什么原因| 什么头什么尾| bossini是什么牌子| 尿蛋白三个加号吃什么药| 为什么腋下有异味| 脚疼是什么原因| 梦到和死人说话是什么意思| 618是什么意思| 为什么会得肾结石| 加码是什么意思| 什么是唐氏儿| 排斥一个人什么意思| 25分贝相当于什么声音| 吃什么可以丰胸| 1901年属什么生肖| hf医学上是什么意思| 童话故事有什么| 什么叫自然拼读| 乌鸦飞进家里什么征兆| 同比和环比是什么意思| 做牛排用什么部位的牛肉| 什么是鸡冠油| 追什么| 异常脑电图说明什么| 孕酮低跟什么有关系| 龙虾吃什么食物| 南瓜不能和什么一起吃| 什么药治咳嗽最好| 侍妾是什么意思| 无创低风险是什么意思| 德国是什么人种| 孟姜女姓什么| 18岁属什么的生肖| 拉黑一个人意味着什么| 会厌炎吃什么药最有效| 乙肝是什么症状| 什么叫集体户口| 什么是天赋| 磅礴是什么意思| 山竹有什么功效和作用| 二月七号是什么星座| nt是什么币| 夏天穿什么鞋子| 情非得已是什么生肖| 尖嘴猴腮什么意思| 今晚吃什么家常菜| 为什么叫印度三哥| 最快的减肥运动是什么| 2025是什么年| 生命科学专业学什么| 激光脱毛对人体有没有什么危害| 纳囊是什么妇科病| 你的书包里有什么英文| 大小脸挂什么科| 五月一号什么星座| 一望无际是什么意思| 测血糖挂号挂什么科| 口头禅是什么意思| 房颤什么意思| 女娲是一个什么样的人| 什么样的伤口需要缝针| media是什么意思| 喝什么茶降血压| 2024什么年| 12月27号是什么星座| 西安有什么| 秦五行属什么| 红豆相思是什么动物| 感冒咳嗽吃什么食物好| brat什么意思| 屁股上有痣代表什么| 喝牛奶就拉肚子是什么原因| 梦到老房子是什么意思| 小肚右边疼是什么原因| 平仓什么意思| 急性肠胃炎吃什么药效果好| 本科生是什么意思| 茄子与什么相克| 日本人什么时候投降的| 感冒后咳嗽吃什么药| 早早孕有什么征兆| 鸡奸什么意思| 下边瘙痒是什么原因| 96年属什么的| 腹部胀气是什么原因| 眼睛出血是什么原因| 坚果什么时候吃最好| 6月6什么星座| 清华大学校长是什么级别| 幼儿园转学需要什么手续| 胃部彩超能检查出什么| 右腿麻木是什么征兆| 喝蛋白粉有什么副作用| 解表药是什么意思| 2024年是什么年| 08属什么生肖| 小孩瘦小不长肉是什么原因| Continental什么牌子| 中元节注意什么| 为什么鼻子无缘无故流鼻血| 心气虚吃什么食物补| 1988年属什么| 5月1日什么星座| 羽毛球拍磅数是什么意思| 北顶娘娘庙求什么灵验| 什么是性瘾| 肠胃不好吃什么菜比较好| 来姨妈为什么是黑色的血| 种植牙有什么风险和后遗症| 吃什么可以减肥肚子| 百度Jump to content

什么叫骨折

From Wikipedia, the free encyclopedia
A visualisation of a solution to the two-dimensional heat equation with temperature represented by the vertical direction and color.
百度 濒临消失的古村落中国古村落之殇10年消失90万个如今,我国传统村落整体上呈现南多北少、东多西少,集中分布于西南、华东地区,云南、贵州两省数量最多。

In mathematics, a partial differential equation (PDE) is an equation which involves a multivariable function and one or more of its partial derivatives.

The function is often thought of as an "unknown" that solves the equation, similar to how x is thought of as an unknown number solving, e.g., an algebraic equation like x2 ? 3x + 2 = 0. However, it is usually impossible to write down explicit formulae for solutions of partial differential equations. There is correspondingly a vast amount of modern mathematical and scientific research on methods to numerically approximate solutions of certain partial differential equations using computers. Partial differential equations also occupy a large sector of pure mathematical research, in which the usual questions are, broadly speaking, on the identification of general qualitative features of solutions of various partial differential equations, such as existence, uniqueness, regularity and stability.[1] Among the many open questions are the existence and smoothness of solutions to the Navier–Stokes equations, named as one of the Millennium Prize Problems in 2000.

Partial differential equations are ubiquitous in mathematically oriented scientific fields, such as physics and engineering. For instance, they are foundational in the modern scientific understanding of sound, heat, diffusion, electrostatics, electrodynamics, thermodynamics, fluid dynamics, elasticity, general relativity, and quantum mechanics (Schr?dinger equation, Pauli equation etc.). They also arise from many purely mathematical considerations, such as differential geometry and the calculus of variations; among other notable applications, they are the fundamental tool in the proof of the Poincaré conjecture from geometric topology.

Partly due to this variety of sources, there is a wide spectrum of different types of partial differential equations, where the meaning of a solution depends on the context of the problem, and methods have been developed for dealing with many of the individual equations which arise. As such, it is usually acknowledged that there is no "universal theory" of partial differential equations, with specialist knowledge being somewhat divided between several essentially distinct subfields.[2]

Ordinary differential equations can be viewed as a subclass of partial differential equations, corresponding to functions of a single variable. Stochastic partial differential equations and nonlocal equations are, as of 2020, particularly widely studied extensions of the "PDE" notion. More classical topics, on which there is still much active research, include elliptic and parabolic partial differential equations, fluid mechanics, Boltzmann equations, and dispersive partial differential equations.[3]

Introduction

[edit]

A function u(x, y, z) of three variables is "harmonic" or "a solution of the Laplace equation" if it satisfies the condition Such functions were widely studied in the 19th century due to their relevance for classical mechanics, for example the equilibrium temperature distribution of a homogeneous solid is a harmonic function. If explicitly given a function, it is usually a matter of straightforward computation to check whether or not it is harmonic. For instance and are both harmonic while is not. It may be surprising that the two examples of harmonic functions are of such strikingly different form. This is a reflection of the fact that they are not, in any immediate way, special cases of a "general solution formula" of the Laplace equation. This is in striking contrast to the case of ordinary differential equations (ODEs) roughly similar to the Laplace equation, with the aim of many introductory textbooks being to find algorithms leading to general solution formulas. For the Laplace equation, as for a large number of partial differential equations, such solution formulas fail to exist.

The nature of this failure can be seen more concretely in the case of the following PDE: for a function v(x, y) of two variables, consider the equation It can be directly checked that any function v of the form v(x, y) = f(x) + g(y), for any single-variable functions f and g whatsoever, will satisfy this condition. This is far beyond the choices available in ODE solution formulas, which typically allow the free choice of some numbers. In the study of PDEs, one generally has the free choice of functions.

The nature of this choice varies from PDE to PDE. To understand it for any given equation, existence and uniqueness theorems are usually important organizational principles. In many introductory textbooks, the role of existence and uniqueness theorems for ODE can be somewhat opaque; the existence half is usually unnecessary, since one can directly check any proposed solution formula, while the uniqueness half is often only present in the background in order to ensure that a proposed solution formula is as general as possible. By contrast, for PDE, existence and uniqueness theorems are often the only means by which one can navigate through the plethora of different solutions at hand. For this reason, they are also fundamental when carrying out a purely numerical simulation, as one must have an understanding of what data is to be prescribed by the user and what is to be left to the computer to calculate.

To discuss such existence and uniqueness theorems, it is necessary to be precise about the domain of the "unknown function". Otherwise, speaking only in terms such as "a function of two variables", it is impossible to meaningfully formulate the results. That is, the domain of the unknown function must be regarded as part of the structure of the PDE itself.

The following provides two classic examples of such existence and uniqueness theorems. Even though the two PDE in question are so similar, there is a striking difference in behavior: for the first PDE, one has the free prescription of a single function, while for the second PDE, one has the free prescription of two functions.

  • Let B denote the unit-radius disk around the origin in the plane. For any continuous function U on the unit circle, there is exactly one function u on B such that and whose restriction to the unit circle is given by U.
  • For any functions f and g on the real line R, there is exactly one function u on R × (?1, 1) such that and with u(x, 0) = f(x) and ??u/?y?(x, 0) = g(x) for all values of x.

Even more phenomena are possible. For instance, the following PDE, arising naturally in the field of differential geometry, illustrates an example where there is a simple and completely explicit solution formula, but with the free choice of only three numbers and not even one function.

  • If u is a function on R2 with then there are numbers a, b, and c with u(x, y) = ax + by + c.

In contrast to the earlier examples, this PDE is nonlinear, owing to the square roots and the squares. A linear PDE is one such that, if it is homogeneous, the sum of any two solutions is also a solution, and any constant multiple of any solution is also a solution.

Definition

[edit]

A partial differential equation is an equation that involves an unknown function of variables and (some of) its partial derivatives.[4] That is, for the unknown function of variables belonging to the open subset of , the -order partial differential equation is defined as where and is the partial derivative operator.

Notation

[edit]

When writing PDEs, it is common to denote partial derivatives using subscripts. For example: In the general situation that u is a function of n variables, then ui denotes the first partial derivative relative to the i-th input, uij denotes the second partial derivative relative to the i-th and j-th inputs, and so on.

The Greek letter Δ denotes the Laplace operator; if u is a function of n variables, then In the physics literature, the Laplace operator is often denoted by ?2; in the mathematics literature, ?2u may also denote the Hessian matrix of u.

Classification

[edit]

Linear and nonlinear equations

[edit]

A PDE is called linear if it is linear in the unknown and its derivatives. For example, for a function u of x and y, a second order linear PDE is of the form where ai and f are functions of the independent variables x and y only. (Often the mixed-partial derivatives uxy and uyx will be equated, but this is not required for the discussion of linearity.) If the ai are constants (independent of x and y) then the PDE is called linear with constant coefficients. If f is zero everywhere then the linear PDE is homogeneous, otherwise it is inhomogeneous. (This is separate from asymptotic homogenization, which studies the effects of high-frequency oscillations in the coefficients upon solutions to PDEs.)

Nearest to linear PDEs are semi-linear PDEs, where only the highest order derivatives appear as linear terms, with coefficients that are functions of the independent variables. The lower order derivatives and the unknown function may appear arbitrarily. For example, a general second order semi-linear PDE in two variables is

In a quasilinear PDE the highest order derivatives likewise appear only as linear terms, but with coefficients possibly functions of the unknown and lower-order derivatives: Many of the fundamental PDEs in physics are quasilinear, such as the Einstein equations of general relativity and the Navier–Stokes equations describing fluid motion.

A PDE without any linearity properties is called fully nonlinear, and possesses nonlinearities on one or more of the highest-order derivatives. An example is the Monge–Ampère equation, which arises in differential geometry.[5]

Second order equations

[edit]

The elliptic/parabolic/hyperbolic classification provides a guide to appropriate initial- and boundary conditions and to the smoothness of the solutions. Assuming uxy = uyx, the general linear second-order PDE in two independent variables has the form where the coefficients A, B, C... may depend upon x and y. If A2 + B2 + C2 > 0 over a region of the xy-plane, the PDE is second-order in that region. This form is analogous to the equation for a conic section:

More precisely, replacing ?x by X, and likewise for other variables (formally this is done by a Fourier transform), converts a constant-coefficient PDE into a polynomial of the same degree, with the terms of the highest degree (a homogeneous polynomial, here a quadratic form) being most significant for the classification.

Just as one classifies conic sections and quadratic forms into parabolic, hyperbolic, and elliptic based on the discriminant B2 ? 4AC, the same can be done for a second-order PDE at a given point. However, the discriminant in a PDE is given by B2 ? AC due to the convention of the xy term being 2B rather than B; formally, the discriminant (of the associated quadratic form) is (2B)2 ? 4AC = 4(B2 ? AC), with the factor of 4 dropped for simplicity.

  1. B2 ? AC < 0 (elliptic partial differential equation): Solutions of elliptic PDEs are as smooth as the coefficients allow, within the interior of the region where the equation and solutions are defined. For example, solutions of Laplace's equation are analytic within the domain where they are defined, but solutions may assume boundary values that are not smooth. The motion of a fluid at subsonic speeds can be approximated with elliptic PDEs, and the Euler–Tricomi equation is elliptic where x < 0. By change of variables, the equation can always be expressed in the form: where x and y correspond to changed variables. This justifies Laplace equation as an example of this type.[6]
  2. B2 ? AC = 0 (parabolic partial differential equation): Equations that are parabolic at every point can be transformed into a form analogous to the heat equation by a change of independent variables. Solutions smooth out as the transformed time variable increases. The Euler–Tricomi equation has parabolic type on the line where x = 0. By change of variables, the equation can always be expressed in the form: where x correspond to changed variables. This justifies heat equation, which are of form , as an example of this type.[6]
  3. B2 ? AC > 0 (hyperbolic partial differential equation): hyperbolic equations retain any discontinuities of functions or derivatives in the initial data. An example is the wave equation. The motion of a fluid at supersonic speeds can be approximated with hyperbolic PDEs, and the Euler–Tricomi equation is hyperbolic where x > 0. By change of variables, the equation can always be expressed in the form: where x and y correspond to changed variables. This justifies wave equation as an example of this type.[6]

If there are n independent variables x1, x2 , …, xn, a general linear partial differential equation of second order has the form

The classification depends upon the signature of the eigenvalues of the coefficient matrix ai,j.

  1. Elliptic: the eigenvalues are all positive or all negative.
  2. Parabolic: the eigenvalues are all positive or all negative, except one that is zero.
  3. Hyperbolic: there is only one negative eigenvalue and all the rest are positive, or there is only one positive eigenvalue and all the rest are negative.
  4. Ultrahyperbolic: there is more than one positive eigenvalue and more than one negative eigenvalue, and there are no zero eigenvalues.[7]

The theory of elliptic, parabolic, and hyperbolic equations have been studied for centuries, largely centered around or based upon the standard examples of the Laplace equation, the heat equation, and the wave equation.

However, the classification only depends on linearity of the second-order terms and is therefore applicable to semi- and quasilinear PDEs as well. The basic types also extend to hybrids such as the Euler–Tricomi equation; varying from elliptic to hyperbolic for different regions of the domain, as well as higher-order PDEs, but such knowledge is more specialized.

Systems of first-order equations and characteristic surfaces

[edit]

The classification of partial differential equations can be extended to systems of first-order equations, where the unknown u is now a vector with m components, and the coefficient matrices Aν are m by m matrices for ν = 1, 2, …, n. The partial differential equation takes the form where the coefficient matrices Aν and the vector B may depend upon x and u. If a hypersurface S is given in the implicit form where φ has a non-zero gradient, then S is a characteristic surface for the operator L at a given point if the characteristic form vanishes:

The geometric interpretation of this condition is as follows: if data for u are prescribed on the surface S, then it may be possible to determine the normal derivative of u on S from the differential equation. If the data on S and the differential equation determine the normal derivative of u on S, then S is non-characteristic. If the data on S and the differential equation do not determine the normal derivative of u on S, then the surface is characteristic, and the differential equation restricts the data on S: the differential equation is internal to S.

  1. A first-order system Lu = 0 is elliptic if no surface is characteristic for L: the values of u on S and the differential equation always determine the normal derivative of u on S.
  2. A first-order system is hyperbolic at a point if there is a spacelike surface S with normal ξ at that point. This means that, given any non-trivial vector η orthogonal to ξ, and a scalar multiplier λ, the equation Q(λξ + η) = 0 has m real roots λ1, λ2, …, λm. The system is strictly hyperbolic if these roots are always distinct. The geometrical interpretation of this condition is as follows: the characteristic form Q(ζ) = 0 defines a cone (the normal cone) with homogeneous coordinates ζ. In the hyperbolic case, this cone has nm sheets, and the axis ζ = λξ runs inside these sheets: it does not intersect any of them. But when displaced from the origin by η, this axis intersects every sheet. In the elliptic case, the normal cone has no real sheets.

Analytical solutions

[edit]

Separation of variables

[edit]

Linear PDEs can be reduced to systems of ordinary differential equations by the important technique of separation of variables. This technique rests on a feature of solutions to differential equations: if one can find any solution that solves the equation and satisfies the boundary conditions, then it is the solution (this also applies to ODEs). We assume as an ansatz that the dependence of a solution on the parameters space and time can be written as a product of terms that each depend on a single parameter, and then see if this can be made to solve the problem.[8]

In the method of separation of variables, one reduces a PDE to a PDE in fewer variables, which is an ordinary differential equation if in one variable – these are in turn easier to solve.

This is possible for simple PDEs, which are called separable partial differential equations, and the domain is generally a rectangle (a product of intervals). Separable PDEs correspond to diagonal matrices – thinking of "the value for fixed x" as a coordinate, each coordinate can be understood separately.

This generalizes to the method of characteristics, and is also used in integral transforms.

Method of characteristics

[edit]

The characteristic surface in n = 2-dimensional space is called a characteristic curve.[9] In special cases, one can find characteristic curves on which the first-order PDE reduces to an ODE – changing coordinates in the domain to straighten these curves allows separation of variables, and is called the method of characteristics.

More generally, applying the method to first-order PDEs in higher dimensions, one may find characteristic surfaces.

Integral transform

[edit]

An integral transform may transform the PDE to a simpler one, in particular, a separable PDE. This corresponds to diagonalizing an operator.

An important example of this is Fourier analysis, which diagonalizes the heat equation using the eigenbasis of sinusoidal waves.

If the domain is finite or periodic, an infinite sum of solutions such as a Fourier series is appropriate, but an integral of solutions such as a Fourier integral is generally required for infinite domains. The solution for a point source for the heat equation given above is an example of the use of a Fourier integral.

Change of variables

[edit]

Often a PDE can be reduced to a simpler form with a known solution by a suitable change of variables. For example, the Black–Scholes equation is reducible to the heat equation by the change of variables[10]

Fundamental solution

[edit]

Inhomogeneous equations[clarification needed] can often be solved (for constant coefficient PDEs, always be solved) by finding the fundamental solution (the solution for a point source ), then taking the convolution with the boundary conditions to get the solution.

This is analogous in signal processing to understanding a filter by its impulse response.

Superposition principle

[edit]

The superposition principle applies to any linear system, including linear systems of PDEs. A common visualization of this concept is the interaction of two waves in phase being combined to result in a greater amplitude, for example sin x + sin x = 2 sin x. The same principle can be observed in PDEs where the solutions may be real or complex and additive. If u1 and u2 are solutions of linear PDE in some function space R, then u = c1u1 + c2u2 with any constants c1 and c2 are also a solution of that PDE in the same function space.

Methods for non-linear equations

[edit]

There are no generally applicable analytical methods to solve nonlinear PDEs. Still, existence and uniqueness results (such as the Cauchy–Kowalevski theorem) are often possible, as are proofs of important qualitative and quantitative properties of solutions (getting these results is a major part of analysis).

Nevertheless, some techniques can be used for several types of equations. The h-principle is the most powerful method to solve underdetermined equations. The Riquier–Janet theory is an effective method for obtaining information about many analytic overdetermined systems.

The method of characteristics can be used in some very special cases to solve nonlinear partial differential equations.[11]

In some cases, a PDE can be solved via perturbation analysis in which the solution is considered to be a correction to an equation with a known solution. Alternatives are numerical analysis techniques from simple finite difference schemes to the more mature multigrid and finite element methods. Many interesting problems in science and engineering are solved in this way using computers, sometimes high performance supercomputers.

Lie group method

[edit]

From 1870 Sophus Lie's work put the theory of differential equations on a more satisfactory foundation. He showed that the integration theories of the older mathematicians can, by the introduction of what are now called Lie groups, be referred, to a common source; and that ordinary differential equations which admit the same infinitesimal transformations present comparable difficulties of integration. He also emphasized the subject of transformations of contact.

A general approach to solving PDEs uses the symmetry property of differential equations, the continuous infinitesimal transformations of solutions to solutions (Lie theory). Continuous group theory, Lie algebras and differential geometry are used to understand the structure of linear and nonlinear partial differential equations for generating integrable equations, to find its Lax pairs, recursion operators, B?cklund transform and finally finding exact analytic solutions to the PDE.

Symmetry methods have been recognized to study differential equations arising in mathematics, physics, engineering, and many other disciplines.

Semi-analytical methods

[edit]

The Adomian decomposition method,[12] the Lyapunov artificial small parameter method, and his homotopy perturbation method are all special cases of the more general homotopy analysis method.[13] These are series expansion methods, and except for the Lyapunov method, are independent of small physical parameters as compared to the well known perturbation theory, thus giving these methods greater flexibility and solution generality.

Numerical solutions

[edit]

The three most widely used numerical methods to solve PDEs are the finite element method (FEM), finite volume methods (FVM) and finite difference methods (FDM), as well other kind of methods called meshfree methods, which were made to solve problems where the aforementioned methods are limited. The FEM has a prominent position among these methods and especially its exceptionally efficient higher-order version hp-FEM. Other hybrid versions of FEM and Meshfree methods include the generalized finite element method (GFEM), extended finite element method (XFEM), spectral finite element method (SFEM), meshfree finite element method, discontinuous Galerkin finite element method (DGFEM), element-free Galerkin method (EFGM), interpolating element-free Galerkin method (IEFGM), etc.

Finite element method

[edit]

The finite element method (FEM) (its practical application often known as finite element analysis (FEA)) is a numerical technique for approximating solutions of partial differential equations (PDE) as well as of integral equations using a finite set of functions.[14][15] The solution approach is based either on eliminating the differential equation completely (steady state problems), or rendering the PDE into an approximating system of ordinary differential equations, which are then numerically integrated using standard techniques such as Euler's method, Runge–Kutta, etc.

Finite difference method

[edit]

Finite-difference methods are numerical methods for approximating the solutions to differential equations using finite difference equations to approximate derivatives.

Finite volume method

[edit]

Similar to the finite difference method or finite element method, values are calculated at discrete places on a meshed geometry. "Finite volume" refers to the small volume surrounding each node point on a mesh. In the finite volume method, surface integrals in a partial differential equation that contain a divergence term are converted to volume integrals, using the divergence theorem. These terms are then evaluated as fluxes at the surfaces of each finite volume. Because the flux entering a given volume is identical to that leaving the adjacent volume, these methods conserve mass by design.

Neural networks

[edit]
Physics informed neural networks have been used to solve partial differential equations in both forward and inverse problems in a data driven manner.[16] One example is the reconstructing fluid flow governed by the Navier-Stokes equations. Using physics informed neural networks does not require the often expensive mesh generation that conventional CFD methods rely on.[17][18]

Weak solutions

[edit]

Weak solutions are functions that satisfy the PDE, yet in other meanings than regular sense. The meaning for this term may differ with context, and one of the most commonly used definitions is based on the notion of distributions.

An example[19] for the definition of a weak solution is as follows:

Consider the boundary-value problem given by: where denotes a second-order partial differential operator in divergence form.

We say a is a weak solution if for every , which can be derived by a formal integral by parts.

An example for a weak solution is as follows: is a weak solution satisfying in distributional sense, as formally,

Theoretical Studies

[edit]

As a branch of pure mathematics, the theoretical studies of PDEs focus on the criteria for a solution to exist, the properties of a solution, and finding its formula is often secondary.

Well-posedness

[edit]

Well-posedness refers to a common schematic package of information about a PDE. To say that a PDE is well-posed, one must have:

  • an existence and uniqueness theorem, asserting that by the prescription of some freely chosen functions, one can single out one specific solution of the PDE
  • by continuously changing the free choices, one continuously changes the corresponding solution

This is, by the necessity of being applicable to several different PDE, somewhat vague. The requirement of "continuity", in particular, is ambiguous, since there are usually many inequivalent means by which it can be rigorously defined. It is, however, somewhat unusual to study a PDE without specifying a way in which it is well-posed.

Regularity

[edit]

Regularity refers to the integrability and differentiability of weak solutions, which can often be represented by Sobolev spaces.

This problem arise due to the difficulty in searching for classical solutions. Researchers often tend to find weak solutions at first and then find out whether it is smooth enough to be qualified as a classical solution.

Results from functional analysis are often used in this field of study.

See also

[edit]

Some common PDEs

Types of boundary conditions

Various topics

Notes

[edit]
  1. ^ "Regularity and singularities in elliptic PDE's: beyond monotonicity formulas | EllipticPDE Project | Fact Sheet | H2020". CORDIS | European Commission. Retrieved 2025-08-06.
  2. ^ Klainerman, Sergiu (2010). "PDE as a Unified Subject". In Alon, N.; Bourgain, J.; Connes, A.; Gromov, M.; Milman, V. (eds.). Visions in Mathematics. Modern Birkh?user Classics. Basel: Birkh?user. pp. 279–315. doi:10.1007/978-3-0346-0422-2_10. ISBN 978-3-0346-0421-5.
  3. ^ Erdo?an, M. Burak; Tzirakis, Nikolaos (2016). Dispersive Partial Differential Equations: Wellposedness and Applications. London Mathematical Society Student Texts. Cambridge: Cambridge University Press. ISBN 978-1-107-14904-5.
  4. ^ Evans 1998, pp. 1–2.
  5. ^ Klainerman, Sergiu (2008), "Partial Differential Equations", in Gowers, Timothy; Barrow-Green, June; Leader, Imre (eds.), The Princeton Companion to Mathematics, Princeton University Press, pp. 455–483
  6. ^ a b c Levandosky, Julie. "Classification of Second-Order Equations" (PDF).
  7. ^ Courant and Hilbert (1962), p.182.
  8. ^ Gershenfeld, Neil (2000). The nature of mathematical modeling (Reprinted (with corr.) ed.). Cambridge: Cambridge University Press. p. 27. ISBN 0521570956.
  9. ^ Zachmanoglou & Thoe 1986, pp. 115–116.
  10. ^ Wilmott, Paul; Howison, Sam; Dewynne, Jeff (1995). The Mathematics of Financial Derivatives. Cambridge University Press. pp. 76–81. ISBN 0-521-49789-2.
  11. ^ Logan, J. David (1994). "First Order Equations and Characteristics". An Introduction to Nonlinear Partial Differential Equations. New York: John Wiley & Sons. pp. 51–79. ISBN 0-471-59916-6.
  12. ^ Adomian, G. (1994). Solving Frontier problems of Physics: The decomposition method. Kluwer Academic Publishers. ISBN 9789401582896.
  13. ^ Liao, S. J. (2003). Beyond Perturbation: Introduction to the Homotopy Analysis Method. Boca Raton: Chapman & Hall/ CRC Press. ISBN 1-58488-407-X.
  14. ^ Solin, P. (2005). Partial Differential Equations and the Finite Element Method. Hoboken, New Jersey: J. Wiley & Sons. ISBN 0-471-72070-4.
  15. ^ Solin, P.; Segeth, K. & Dolezel, I. (2003). Higher-Order Finite Element Methods. Boca Raton: Chapman & Hall/CRC Press. ISBN 1-58488-438-X.
  16. ^ Raissi, M.; Perdikaris, P.; Karniadakis, G. E. (2025-08-06). "Physics-informed neural networks: A deep learning framework for solving forward and inverse problems involving nonlinear partial differential equations". Journal of Computational Physics. 378: 686–707. Bibcode:2019JCoPh.378..686R. doi:10.1016/j.jcp.2018.10.045. ISSN 0021-9991. OSTI 1595805. S2CID 57379996.
  17. ^ Mao, Zhiping; Jagtap, Ameya D.; Karniadakis, George Em (2025-08-06). "Physics-informed neural networks for high-speed flows". Computer Methods in Applied Mechanics and Engineering. 360 112789. Bibcode:2020CMAME.360k2789M. doi:10.1016/j.cma.2019.112789. ISSN 0045-7825. S2CID 212755458.
  18. ^ Raissi, Maziar; Yazdani, Alireza; Karniadakis, George Em (2025-08-06). "Hidden fluid mechanics: Learning velocity and pressure fields from flow visualizations". Science. 367 (6481): 1026–1030. Bibcode:2020Sci...367.1026R. doi:10.1126/science.aaw4741. PMC 7219083. PMID 32001523.
  19. ^ Evans 1998, chpt. 6. Second-Order Elliptic Equations.

References

[edit]

Further reading

[edit]
[edit]
扁桃体炎吃什么药最好 皮肤黑穿什么颜色的衣服显白 一什么车厢 气什么意思 淋球菌是什么病
活水是什么意思 抑郁到什么程度要吃氟西汀 uw是什么单位 内分泌紊乱是什么意思 奥运五环绿色代表什么
糖尿病吃什么主食 高血糖有什么症状 斑鸠是什么意思 易经的易是什么意思 更年期有什么症状
饱不洗头饿不洗澡是为什么 什么木头的菜板最好 每个月月经都推迟是什么原因 补睾丸吃什么药最好 人生得意须尽欢什么意思
拔智齿后吃什么hcv9jop2ns6r.cn 桃花什么时候开放hcv9jop7ns3r.cn vd是什么意思hcv9jop6ns3r.cn 伟哥是什么意思hcv8jop8ns6r.cn 雷锋日是什么时候hcv9jop1ns3r.cn
脂肪瘤挂什么科hcv8jop2ns5r.cn ag是什么意思hcv8jop0ns6r.cn 孩子注意力不集中是什么原因bysq.com 烤冷面的面皮是什么面hcv7jop6ns8r.cn 怀孕后为什么会生化hcv8jop4ns4r.cn
婚检男性检查什么hcv8jop8ns9r.cn 石蜡是什么东西sscsqa.com epl是什么意思hcv7jop6ns6r.cn 五光十色是什么生肖hcv7jop9ns7r.cn kaiser是什么品牌hcv8jop0ns4r.cn
圆明园是什么时候被烧的hcv8jop1ns5r.cn 井泉水命什么意思hcv8jop6ns3r.cn 炒作是什么意思hcv9jop4ns2r.cn 为什么会起水泡hcv8jop2ns4r.cn 毋庸置疑什么意思hcv7jop6ns5r.cn
百度