铁是什么颜色| 王字旁一个行念什么| 胆红素三个指标都高是什么原因| 为什么每天晚上睡觉都做梦| 心律不齐吃什么食物好| 神经性耳鸣吃什么药| 芸字五行属什么| 甲状腺在什么位置| 为什么蚊子不咬我| 蜂鸟是什么鸟| 法大大是什么| fl是什么| 头痛是什么原因造成的| 副主任科员是什么级别| 舌炎吃什么药效果最好| 卵泡是什么意思| 喝咖啡对身体有什么好处| 撮鸟是什么意思| 2021年是什么年| 基础代谢是什么意思| 球蛋白高是什么意思| 马齿苋长什么样子| 奇变偶不变是什么意思| 凉皮用什么面粉| 医生为为什么建议不吃生菜| 卧轨是什么意思| 3.1号是什么星座| 风生水起是什么生肖| 田五行属性是什么| 减肥期间适合吃什么| 拉伤筋用什么药好| 红蓝是什么意思| 前列腺炎是什么意思| 趾高气昂是什么意思| 镇党委副书记是什么级别| 凶猛的动物是什么生肖| 一个三点水一个令念什么| 肝内低密度灶是什么意思| 灰色配什么颜色好看| 血糖高适合吃什么水果| 哎呀是什么意思| 每天吃松子有什么好处| 头头是道什么意思| 喝酒后头晕是什么原因| 消炎药不能和什么一起吃| 右眼皮一直跳是什么预兆| 脱水是什么意思| 屁股流血是什么原因| 紧急避孕药叫什么名字| 生孩子前要注意什么| 什么饺子馅最好吃| 定坤丹适合什么人吃| 石英表是什么意思| 金价下跌意味着什么| 尖锐湿疣什么症状| 刺史相当于现在的什么官| 什么情况下需要安装心脏起搏器| 例假少吃什么药| 生姜和红糖熬水有什么作用| 扁平疣是什么原因造成的| 胃炎是什么原因引起的| 宝贝是什么意思| 瘤是什么意思| 枕头发黄是什么原因| 什么是夫妻宫| 农历五月二十四是什么星座| 杀青原指什么| 强劲的动物是什么生肖| 做梦和别人吵架意味着什么| mirage轮胎什么牌子| 为什么夏天容易拉肚子| 女人是男人的什么| 视力5.3是什么概念| nk细胞是什么| 宫保鸡丁是什么菜系| 嫦娥是什么生肖| 红斑狼疮什么症状| 肛裂是什么原因造成的| 发烧感冒吃什么药| 心脏病吃什么水果最好| 蜈蚣怕什么| 红茶什么季节喝最好| 炙是什么意思| 1217是什么星座| 百合长什么样子| 桂枝和肉桂有什么区别| 呼和浩特有什么特产| 什么是酮体| 嫩模是什么意思| 脚底出汗是什么原因女| fnc是什么意思| 鲔鱼是什么鱼| 细菌感染吃什么药好| 牙冠什么材质的好| 取活检是什么意思| 支付宝提现是什么意思| lcp是什么意思| 藩王是什么意思| 牙齿涂氟是什么意思| 农历五月初五是什么节| sk是什么| 喝茶对人体有什么好处| 山炮是什么意思| 洁身自爱是什么生肖| 什么人容易得帕金森病| 什么的芦花| 宝宝惊跳反射什么时候消失| 国企混改是什么意思| 茵陈是什么植物| 新生儿呛奶是什么原因引起的| 慢性盆腔炎吃什么药| 吃什么变碱性体质最快| 相宜的意思是什么| 铁饱和度低什么原因| 抗核小体抗体阳性说明什么| 什么是南红| 什么是苔藓皮肤病| 劳燕分飞是什么意思| 晚上七点是什么时辰| 草字头弓读什么字| 眼底筛查是检查什么| 利郎男装是什么档次的| 什么药可以降肌酐| 公安厅长是什么级别| 滑膜炎吃什么好得快| 呼吸困难是什么原因| 风雨交加是什么生肖| 公丁香和母丁香有什么区别| 二聚体测定是什么| 海参什么季节吃好| 安陵容什么时候变坏的| msgm是什么品牌| 柠檬水有什么好处| 什么是创业板股票| fierce是什么意思| 鸟字旁与什么有关| 心脏右束支传导阻滞是什么意思| 女生右眼睛老是跳是什么原因| 台风什么时候结束| 洋葱为什么会让人流泪| 舌苔发白是什么原因引起的| 散光是什么原因造成的| 夜半是什么时辰| 十八层地狱分别叫什么| 小孩长白头发是什么原因| 大姨的女儿叫什么| 血压低有什么危害| 山川载不动太多悲哀是什么歌| 乔迁之喜送什么| 孙权和孙策是什么关系| 清热去湿热颗粒有什么功效| 床虱咬了要擦什么药膏| 葫芦是什么生肖| 体检前一天晚上吃什么| 阴唇为什么会变黑| 为什么当兵| 三阳开泰是什么意思| 梦到自己掉头发是什么预兆| 包装饮用水是什么水| 作恶多端是什么意思| 吃完杏不能吃什么| 宝宝胎动频繁说明什么| 吃什么食物可以降尿酸| 全飞秒手术是什么| 什么叫双规| 156是什么意思| 学分是什么意思| 镇长属于什么级别| 后背长痘痘用什么药膏| 酒精对皮肤有什么伤害| 吃什么中药能降血压| 毛囊炎什么症状| 痔疮吃什么药好| 身份证前六位代表什么| 小腿发胀是什么原因| 吃什么食物最补血| 葛根长什么样子图片| 轻度脑梗吃什么药最好| 血型o型rh阳性是什么意思| 惊讶的什么| 09年属什么生肖| 血管炎是什么病| 孕妇梦见老公出轨是什么意思| 什么情况下吃奥司他韦| 荷叶茶有什么功效和作用| 投诉医生打什么电话| 回族不能吃什么| 为什么吃了避孕药还是怀孕了| 居高临下的临是什么意思| 牙齿根管治疗是什么意思| 什么样的人爱长结节| 朱元璋为什么不杀朱棣| 坚强后盾是什么意思| lr是什么| 什么什么不周| 胎动突然频繁是什么原因| 腹泻拉水吃什么药| 多潘立酮片是什么药| 嗓子疼是什么原因引起的| 梦见种地是什么意思| 吃什么对肺最好| 咳嗽有痰挂什么科| 枸杞什么季节成熟| 不老实是什么意思| 什么牌子的益生菌调理肠胃比较好| 立flag什么意思| 种植什么最赚钱农村| 2009年属什么| 家里为什么有蜈蚣| 小孩牙疼吃什么药| 风流人物指什么生肖| 减胎对另一个胎儿有什么影响| 离线缓存是什么意思| 近视是什么意思| 湖南湖北以什么湖为界| 心脏ct能检查出什么| ab型血可以输什么血| 群什么吐什么| 臣附议是什么意思| 圆珠笔是什么笔| 什么是水肿| 33数字代表什么意思| 小猫打什么疫苗| 小鸭子吃什么食物| 1887年属什么生肖| 血管紧张素是什么意思| camus是什么酒| 脖子上长个包挂什么科| 同学生日送什么礼物| 学区房什么意思| 去医院看嘴唇挂什么科| 脂肪肝吃什么食物| 心什么什么什么| 粘液丝高是什么原因| 周文王叫什么| 膈是什么器官| 什么人容易得心梗| 鸡炖什么好吃| 角是什么意思| 中国现在是什么社会| 什么蓝牙耳机好| mmc是什么意思| 胃热口干口苦口臭吃什么药好| 同字五行属什么| 小孩子腿疼是什么原因| 盆腔积液是什么原因| 什么东西可以代替阴茎| 尿道口流白色液体是什么病| 用进废退什么意思| 淋巴结肿吃什么消炎药| 挖野菜是什么意思| 纳呆是什么意思| 胆结石切除胆囊后有什么影响| 月破是什么意思| 候车是什么意思| 包公代表什么生肖| 办慢性病有什么好处| 哈衣是什么意思| 下午四点到五点是什么时辰| 什么是直系亲属| 考妣是什么意思| 什么的元帅| 安坦又叫什么药| 五指姑娘是什么意思| 百度Jump to content

新赛季国象甲级联赛4月揭幕

From Wikipedia, the free encyclopedia
The transformation P is the orthogonal projection onto the line m.
百度 沈国明认为,巨石现在出口美国的占比大约为7-8%,单一来看是一个重要的市场,但是还不至于是一个威胁到公司的生死的致命市场。

In linear algebra and functional analysis, a projection is a linear transformation from a vector space to itself (an endomorphism) such that . That is, whenever is applied twice to any vector, it gives the same result as if it were applied once (i.e. is idempotent). It leaves its image unchanged.[1] This definition of "projection" formalizes and generalizes the idea of graphical projection. One can also consider the effect of a projection on a geometrical object by examining the effect of the projection on points in the object.

Definitions

[edit]

A projection on a vector space is a linear operator such that .

When has an inner product and is complete, i.e. when is a Hilbert space, the concept of orthogonality can be used. A projection on a Hilbert space is called an orthogonal projection if it satisfies for all . A projection on a Hilbert space that is not orthogonal is called an oblique projection.

Projection matrix

[edit]
  • A square matrix is called a projection matrix if it is equal to its square, i.e. if .[2]:?p. 38?
  • A square matrix is called an orthogonal projection matrix if for a real matrix, and respectively for a complex matrix, where denotes the transpose of and denotes the adjoint or Hermitian transpose of .[2]:?p. 223?
  • A projection matrix that is not an orthogonal projection matrix is called an oblique projection matrix.

The eigenvalues of a projection matrix must be 0 or 1.

Examples

[edit]

Orthogonal projection

[edit]

For example, the function which maps the point in three-dimensional space to the point is an orthogonal projection onto the xy-plane. This function is represented by the matrix

The action of this matrix on an arbitrary vector is

To see that is indeed a projection, i.e., , we compute

Observing that shows that the projection is an orthogonal projection.

Oblique projection

[edit]

A simple example of a non-orthogonal (oblique) projection is

Via matrix multiplication, one sees that showing that is indeed a projection.

The projection is orthogonal if and only if because only then

Properties and classification

[edit]
The transformation T is the projection along k onto m. The range of T is m and the kernel is k.

Idempotence

[edit]

By definition, a projection is idempotent (i.e. ).

Open map

[edit]

Every projection is an open map onto its image, meaning that it maps each open set in the domain to an open set in the subspace topology of the image.[citation needed] That is, for any vector and any ball (with positive radius) centered on , there exists a ball (with positive radius) centered on that is wholly contained in the image .

Complementarity of image and kernel

[edit]

Let be a finite-dimensional vector space and be a projection on . Suppose the subspaces and are the image and kernel of respectively. Then has the following properties:

  1. is the identity operator on :
  2. We have a direct sum . Every vector may be decomposed uniquely as with and , and where

The image and kernel of a projection are complementary, as are and . The operator is also a projection as the image and kernel of become the kernel and image of and vice versa. We say is a projection along onto (kernel/image) and is a projection along onto .

Spectrum

[edit]

In infinite-dimensional vector spaces, the spectrum of a projection is contained in as Only 0 or 1 can be an eigenvalue of a projection. This implies that an orthogonal projection is always a positive semi-definite matrix. In general, the corresponding eigenspaces are (respectively) the kernel and range of the projection. Decomposition of a vector space into direct sums is not unique. Therefore, given a subspace , there may be many projections whose range (or kernel) is .

If a projection is nontrivial it has minimal polynomial , which factors into distinct linear factors, and thus is diagonalizable.

Product of projections

[edit]

The product of projections is not in general a projection, even if they are orthogonal. If two projections commute then their product is a projection, but the converse is false: the product of two non-commuting projections may be a projection.

If two orthogonal projections commute then their product is an orthogonal projection. If the product of two orthogonal projections is an orthogonal projection, then the two orthogonal projections commute (more generally: two self-adjoint endomorphisms commute if and only if their product is self-adjoint).

Orthogonal projections

[edit]

When the vector space has an inner product and is complete (is a Hilbert space) the concept of orthogonality can be used. An orthogonal projection is a projection for which the range and the kernel are orthogonal subspaces. Thus, for every and in , . Equivalently:

A projection is orthogonal if and only if it is self-adjoint. Using the self-adjoint and idempotent properties of , for any and in we have , , and where is the inner product associated with . Therefore, and are orthogonal projections.[3] The other direction, namely that if is orthogonal then it is self-adjoint, follows from the implication from to for every and in ; thus .

The existence of an orthogonal projection onto a closed subspace follows from the Hilbert projection theorem.

Properties and special cases

[edit]

An orthogonal projection is a bounded operator. This is because for every in the vector space we have, by the Cauchy–Schwarz inequality: Thus .

For finite-dimensional complex or real vector spaces, the standard inner product can be substituted for .

Formulas
[edit]

A simple case occurs when the orthogonal projection is onto a line. If is a unit vector on the line, then the projection is given by the outer product (If is complex-valued, the transpose in the above equation is replaced by a Hermitian transpose). This operator leaves u invariant, and it annihilates all vectors orthogonal to , proving that it is indeed the orthogonal projection onto the line containing u.[4] A simple way to see this is to consider an arbitrary vector as the sum of a component on the line (i.e. the projected vector we seek) and another perpendicular to it, . Applying projection, we get by the properties of the dot product of parallel and perpendicular vectors.

This formula can be generalized to orthogonal projections on a subspace of arbitrary dimension. Let be an orthonormal basis of the subspace , with the assumption that the integer , and let denote the matrix whose columns are , i.e., . Then the projection is given by:[5] which can be rewritten as

The matrix is the partial isometry that vanishes on the orthogonal complement of , and is the isometry that embeds into the underlying vector space. The range of is therefore the final space of . It is also clear that is the identity operator on .

The orthonormality condition can also be dropped. If is a (not necessarily orthonormal) basis with , and is the matrix with these vectors as columns, then the projection is:[6][7]

The matrix still embeds into the underlying vector space but is no longer an isometry in general. The matrix is a "normalizing factor" that recovers the norm. For example, the rank-1 operator is not a projection if After dividing by we obtain the projection onto the subspace spanned by .

In the general case, we can have an arbitrary positive definite matrix defining an inner product , and the projection is given by . Then

When the range space of the projection is generated by a frame (i.e. the number of generators is greater than its dimension), the formula for the projection takes the form: . Here stands for the Moore–Penrose pseudoinverse. This is just one of many ways to construct the projection operator.

If is a non-singular matrix and (i.e., is the null space matrix of ),[8] the following holds:

If the orthogonal condition is enhanced to with non-singular, the following holds:

All these formulas also hold for complex inner product spaces, provided that the conjugate transpose is used instead of the transpose. Further details on sums of projectors can be found in Banerjee and Roy (2014).[9] Also see Banerjee (2004)[10] for application of sums of projectors in basic spherical trigonometry.

Oblique projections

[edit]

The term oblique projections is sometimes used to refer to non-orthogonal projections. These projections are also used to represent spatial figures in two-dimensional drawings (see oblique projection), though not as frequently as orthogonal projections. Whereas calculating the fitted value of an ordinary least squares regression requires an orthogonal projection, calculating the fitted value of an instrumental variables regression requires an oblique projection.

A projection is defined by its kernel and the basis vectors used to characterize its range (which is a complement of the kernel). When these basis vectors are orthogonal to the kernel, then the projection is an orthogonal projection. When these basis vectors are not orthogonal to the kernel, the projection is an oblique projection, or just a projection.

A matrix representation formula for a nonzero projection operator

[edit]

Let be a linear operator such that and assume that is not the zero operator. Let the vectors form a basis for the range of , and assemble these vectors in the matrix . Then , otherwise and is the zero operator. The range and the kernel are complementary spaces, so the kernel has dimension . It follows that the orthogonal complement of the kernel has dimension . Let form a basis for the orthogonal complement of the kernel of the projection, and assemble these vectors in the matrix . Then the projection (with the condition ) is given by

This expression generalizes the formula for orthogonal projections given above.[11][12] A standard proof of this expression is the following. For any vector in the vector space , we can decompose , where vector is in the image of , and vector So , and then is in the kernel of , which is the null space of In other words, the vector is in the column space of so for some dimension vector and the vector satisfies by the construction of . Put these conditions together, and we find a vector so that . Since matrices and are of full rank by their construction, the -matrix is invertible. So the equation gives the vector In this way, for any vector and hence .

In the case that is an orthogonal projection, we can take , and it follows that . By using this formula, one can easily check that . In general, if the vector space is over complex number field, one then uses the Hermitian transpose and has the formula . Recall that one can express the Moore–Penrose inverse of the matrix by since has full column rank, so .

Singular values

[edit]

is also an oblique projection. The singular values of and can be computed by an orthonormal basis of . Let be an orthonormal basis of and let be the orthogonal complement of . Denote the singular values of the matrix by the positive values . With this, the singular values for are:[13] and the singular values for are This implies that the largest singular values of and are equal, and thus that the matrix norm of the oblique projections are the same. However, the condition number satisfies the relation , and is therefore not necessarily equal.

Finding projection with an inner product

[edit]

Let be a vector space (in this case a plane) spanned by orthogonal vectors . Let be a vector. One can define a projection of onto as where repeated indices are summed over (Einstein sum notation). The vector can be written as an orthogonal sum such that . is sometimes denoted as . There is a theorem in linear algebra that states that this is the smallest distance (the orthogonal distance) from to and is commonly used in areas such as machine learning.

y is being projected onto the vector space V.

Canonical forms

[edit]

Any projection on a vector space of dimension over a field is a diagonalizable matrix, since its minimal polynomial divides , which splits into distinct linear factors. Thus there exists a basis in which has the form

where is the rank of . Here is the identity matrix of size , is the zero matrix of size , and is the direct sum operator. If the vector space is complex and equipped with an inner product, then there is an orthonormal basis in which the matrix of P is[14]

where . The integers and the real numbers are uniquely determined. . The factor corresponds to the maximal invariant subspace on which acts as an orthogonal projection (so that P itself is orthogonal if and only if ) and the -blocks correspond to the oblique components.

Projections on normed vector spaces

[edit]

When the underlying vector space is a (not necessarily finite-dimensional) normed vector space, analytic questions, irrelevant in the finite-dimensional case, need to be considered. Assume now is a Banach space.

Many of the algebraic results discussed above survive the passage to this context. A given direct sum decomposition of into complementary subspaces still specifies a projection, and vice versa. If is the direct sum , then the operator defined by is still a projection with range and kernel . It is also clear that . Conversely, if is projection on , i.e. , then it is easily verified that . In other words, is also a projection. The relation implies and is the direct sum .

However, in contrast to the finite-dimensional case, projections need not be continuous in general. If a subspace of is not closed in the norm topology, then the projection onto is not continuous. In other words, the range of a continuous projection must be a closed subspace. Furthermore, the kernel of a continuous projection (in fact, a continuous linear operator in general) is closed. Thus a continuous projection gives a decomposition of into two complementary closed subspaces: .

The converse holds also, with an additional assumption. Suppose is a closed subspace of . If there exists a closed subspace such that X = UV, then the projection with range and kernel is continuous. This follows from the closed graph theorem. Suppose xnx and Pxny. One needs to show that . Since is closed and {Pxn} ? U, y lies in , i.e. Py = y. Also, xn ? Pxn = (I ? P)xnx ? y. Because is closed and {(I ? P)xn} ? V, we have , i.e. , which proves the claim.

The above argument makes use of the assumption that both and are closed. In general, given a closed subspace , there need not exist a complementary closed subspace , although for Hilbert spaces this can always be done by taking the orthogonal complement. For Banach spaces, a one-dimensional subspace always has a closed complementary subspace. This is an immediate consequence of Hahn–Banach theorem. Let be the linear span of . By Hahn–Banach, there exists a bounded linear functional such that φ(u) = 1. The operator satisfies , i.e. it is a projection. Boundedness of implies continuity of and therefore is a closed complementary subspace of .

Applications and further considerations

[edit]

Projections (orthogonal and otherwise) play a major role in algorithms for certain linear algebra problems:

As stated above, projections are a special case of idempotents. Analytically, orthogonal projections are non-commutative generalizations of characteristic functions. Idempotents are used in classifying, for instance, semisimple algebras, while measure theory begins with considering characteristic functions of measurable sets. Therefore, as one can imagine, projections are very often encountered in the context of operator algebras. In particular, a von Neumann algebra is generated by its complete lattice of projections.

Generalizations

[edit]

More generally, given a map between normed vector spaces one can analogously ask for this map to be an isometry on the orthogonal complement of the kernel: that be an isometry (compare Partial isometry); in particular it must be onto. The case of an orthogonal projection is when W is a subspace of V. In Riemannian geometry, this is used in the definition of a Riemannian submersion.

See also

[edit]

Notes

[edit]
  1. ^ Meyer, pp 386+387
  2. ^ a b Horn, Roger A.; Johnson, Charles R. (2013). Matrix Analysis, second edition. Cambridge University Press. ISBN 9780521839402.
  3. ^ Meyer, p. 433
  4. ^ Meyer, p. 431
  5. ^ Meyer, equation (5.13.4)
  6. ^ Banerjee, Sudipto; Roy, Anindya (2014), Linear Algebra and Matrix Analysis for Statistics, Texts in Statistical Science (1st ed.), Chapman and Hall/CRC, ISBN 978-1420095388
  7. ^ Meyer, equation (5.13.3)
  8. ^ See also Linear least squares (mathematics) § Properties of the least-squares estimators.
  9. ^ Banerjee, Sudipto; Roy, Anindya (2014), Linear Algebra and Matrix Analysis for Statistics, Texts in Statistical Science (1st ed.), Chapman and Hall/CRC, ISBN 978-1420095388
  10. ^ Banerjee, Sudipto (2004), "Revisiting Spherical Trigonometry with Orthogonal Projectors", The College Mathematics Journal, 35 (5): 375–381, doi:10.1080/07468342.2004.11922099, S2CID 122277398
  11. ^ Banerjee, Sudipto; Roy, Anindya (2014), Linear Algebra and Matrix Analysis for Statistics, Texts in Statistical Science (1st ed.), Chapman and Hall/CRC, ISBN 978-1420095388
  12. ^ Meyer, equation (7.10.39)
  13. ^ Brust, J. J.; Marcia, R. F.; Petra, C. G. (2020), "Computationally Efficient Decompositions of Oblique Projection Matrices", SIAM Journal on Matrix Analysis and Applications, 41 (2): 852–870, doi:10.1137/19M1288115, OSTI 1680061, S2CID 219921214
  14. ^ Dokovi?, D. ?. (August 1991). "Unitary similarity of projectors". Aequationes Mathematicae. 42 (1): 220–224. doi:10.1007/BF01818492. S2CID 122704926.

References

[edit]
  • Banerjee, Sudipto; Roy, Anindya (2014), Linear Algebra and Matrix Analysis for Statistics, Texts in Statistical Science (1st ed.), Chapman and Hall/CRC, ISBN 978-1420095388
  • Dunford, N.; Schwartz, J. T. (1958). Linear Operators, Part I: General Theory. Interscience.
  • Meyer, Carl D. (2000). Matrix Analysis and Applied Linear Algebra. Society for Industrial and Applied Mathematics. ISBN 978-0-89871-454-8.
  • Brezinski, Claude: Projection Methods for Systems of Equations, North-Holland, ISBN 0-444-82777-3 (1997).
[edit]
人为什么会梦游 吃布洛芬不能吃什么 舌头发热是什么原因 梦见情敌什么预兆 宫外孕术后可以吃什么
ly是什么意思 易烊千玺原名叫什么 无咎是什么意思 南极为什么比北极冷 心服口服是什么意思
什么样的人能镇住凶宅 宝宝什么时候断奶最好 臭虫是什么 双肺纹理增多是什么意思 918是什么星座
预判是什么意思 一面之词是什么意思 男性尿道炎是什么原因引起的 为什么小脑会萎缩 梦见吃肉是什么意思
梦见输钱是什么预兆hcv8jop1ns1r.cn 食欲不振是什么原因hcv7jop7ns4r.cn 鱼完念什么hcv9jop2ns2r.cn 柳树代表什么生肖hcv9jop0ns9r.cn 随性是什么意思hcv8jop3ns7r.cn
手机账号是什么jasonfriends.com 感悟是什么意思qingzhougame.com 沙和尚的武器叫什么hcv8jop3ns7r.cn 女性看乳房应该挂什么科hcv8jop9ns4r.cn 胸口痛吃什么药hcv9jop6ns9r.cn
满月是什么意思hcv8jop0ns4r.cn 女生来大姨妈要注意什么aiwuzhiyu.com 头晕眼睛模糊是什么原因jasonfriends.com 脸黄是什么原因造成的qingzhougame.com 尿液中有血是什么原因onlinewuye.com
七月份可以种什么菜hcv9jop6ns8r.cn 宋朝之前是什么朝代hcv8jop4ns7r.cn 做爱女生什么感觉hcv9jop1ns1r.cn 富氢水是什么hcv9jop5ns7r.cn 前三个月怀孕注意什么hcv8jop3ns3r.cn
百度