Saturday, January 15, 2022

Recent Questions - Mathematics Stack Exchange

Recent Questions - Mathematics Stack Exchange


Smoothing function details in analytic number theory paper

Posted: 15 Jan 2022 09:51 AM PST

My question: How does the bound \[ g_c(l)\ll 1/(xl)^2\hspace {10mm}\text {for }x^{1-2\epsilon }l>c^2\hspace {10mm}(1)\] towards the bottom of page 277 of http://matwbn.icm.edu.pl/ksiazki/aa/aa61/aa6134.pdf work?

How far I can get: Working upwards from the stated bound (you need pages 276 and 277), you can see that \[ g_c(l)=\sum _{l_1l_2=l}\int \int F(uc)F(vc)F(uvc^2)e\left (ul_1+vl_2\right )dudv=\frac {1}{c^2}\sum _{l_1l_2=l}\int \int F(u)F(v)F(uv)e\left (ul_1/c+vl_2/c\right )dudv.\] The function $F$ isn't given explicitly (you need page 276), and maybe that's where I get a bit confused. It seems maybe they're saying that $F'(t)=0$ outside of $(x-x^{1-\epsilon },x)$, and on that interval it is $\ll 1/x^{1-\epsilon }$? And I guess the higher derivatives then satisfy $F^{(n)}(t)\ll 1/x^{n(1-\epsilon )}$? So \[ \left (\frac {d}{dudv}\right )^n\Big \{ F(u)F(v)F(uv)\Big \} \ll \frac {x^n}{x^{2n(1-\epsilon )}}=\frac {1}{x^{n(1-2\epsilon )}}\] Integrating the above integral for $g_c(l)$, it seems to me that by each partial integration (of both variables) we get a factor $c^2/l$, so after $n$ integrations (of both variables) we are left with \[ g_c(l)=\frac {1}{c^2}\left (\frac {c^2}{l}\right )^n\int \int \left (\frac {d}{dudv}\right )^n\Big \{ F(u)F(v)F(uv)\Big \} e(...)\ll \frac {1}{c^2}\left (\frac {c^2}{lx^{1-2\epsilon }}\right )^n\int \int dudv\] but I don't know how to take this further - probably I've gone wrong already.

Can anyone help me fill out the details of (1) correctly?

Regarding the degree of a map $f:S^n\rightarrow S^n$

Posted: 15 Jan 2022 09:47 AM PST

The degree of a map $f:S^n\rightarrow S^n$ is defined as an integer d such that $f_*(\alpha) = d\alpha$ where $\alpha$ generates the infinite cyclic group $H_n(S^n)$. So my question is, in some cases the map $f$ will defined from $S^n$ to a copy of $S^n$. So for degree to be well defined we need a canonical mapping from $S^n$ to all its copies. Now this is essentially what one might call as the identity map. Can someone tell me how this map is defined?

Show $\sum_{i=1}^{n}(\overline{\mathbf{x}}^T\mathbf{u_i})\mathbf{u_i} = \overline{\mathbf{x}}$ component-wise?

Posted: 15 Jan 2022 09:49 AM PST

We have that $\mathbf{u_i}^T\mathbf{u_j} = 1$ when $i=j$ and $0$ otherwise. Then $\mathbf{u_j}^T\sum_{i=1}^{n}(\overline{\mathbf{x}}^T\mathbf{u_i})\mathbf{u_i} = \mathbf{u_j}^T\overline{\mathbf{x}}$ which implies that $\sum_{i=1}^{n}(\overline{\mathbf{x}}^T\mathbf{u_i})\mathbf{u_i} = \overline{\mathbf{x}}$. I have tried to show this component-wise($\overline{\mathbf{x}}_k = (\sum_{i=1}^{n}(\overline{\mathbf{x}}^T\mathbf{u_i})\mathbf{u_i})_k$). However, I cannot end up with the equation. Here is where I started: $$ \overline{\mathbf{x}}_k = \sum_{i=1}^{n} (\sum_{j=1}^{n}{\overline{\mathbf{x}}_j \mathbf{u_i}_j}) \mathbf{u_i}_k $$

What does it mean for a graph to be $P_3$-free (or more generally $P_k$-free)?

Posted: 15 Jan 2022 09:42 AM PST

I'm confused on this. Does it mean that there is no induced subgraph that is $P_3$? Does it mean that any set of 3 connected vertices must be a triangle? Or, for $P_4$, a square?

Pictures would be super helpful if possible!

Quick way to find the number of real/integral solutions of an equation in two variables

Posted: 15 Jan 2022 09:42 AM PST

For Example, the question is $$xy - 6{(x+y)} = 0$$

We need to find the number of integral solutions to this equation, constraint being $x \leq y$

Options :

  1. 5
  2. 10
  3. 12
  4. 9

Trouble understanding the use of the replacement theorem in identifying infinite dimensional vector spaces

Posted: 15 Jan 2022 09:43 AM PST

In chapter 2 of Axler's Linear Algebra Done Right the following theorem is given

In a finite-dimensional vector space, the length of every linearly independent list of vectors is less than or equal to the length of every spanning list of vectors.

Along with the following footnote

Suppose that for each positive integer $m$, there exists a linearly independent list of $m$ vectors in $V$. Then this theorem implies that $V$ is infinite dimensional.

I do not understand what exactly this result states, I have pondered this statement for hours and have looked all over the internet but I could not find an explanation. The closest I have gotten is this old post, but the answer was more focused on the result as an implication of the replacement theorem rather than the meaning of the result itself.

All I can conclude thus far is that: If there is an index $ m \in \mathbb{Z}^+$ such that for each value of $m$ there is a linearly independent list of the form $(a_1,a_2,...,a_m)$ vectors in $V$ then $V$ is infinite dimensional.

Which does not really make sense to me.

Fundamental Theorem of Sequences | Oscillating Sequences

Posted: 15 Jan 2022 09:41 AM PST

I am new to calculus and as I was learning about sequences and limits of sequences I have obviously met the FTS which basically states that any increasing sequences bounded above / any decreasing sequences bounded below all tend to a limit. The theorem sounds very clear to me, though I do not understand how can I relate this theorem to oscillating sequences, which are neither increasing nor decreasing, though some are bounded and have a limit.
Please can somebody explain this to me ( or you could just tell me that I went wrong somewhere in my understanding of FTS)

key: FTS - The Fundamental Theorem of Sequences

The question is what is the maximum likelihood estimator of $\mu_2$? Is it that equal to $\hat{\mu}_2=\frac{1}{n}\sum X_i$?

Posted: 15 Jan 2022 09:30 AM PST

Suppose $X_{1}, X_{2},\ldots,X_{n}$ are i.i.d. p dimension random vectors observations from a multivariate normal distribution $N(\mu,\Sigma)$ where $\Sigma$ is known. Suppose that $\mu=[\mu_1, \mu_2]^T$ and $\mu_1$ with dimension $s$. If we try to use the likelihood ratio procedure to produce a test statistic $H_0: \mu_1=0$ v.s. $H_a: \mu_1 \neq 0$.

The question is what is the maximum likelihood estimator of $\mu_2$? Is it that equal to $\hat{\mu}_2=\frac{1}{n}\sum X_i$?

Finding the distribution of the total time spent in a transient state of a continious time Markov chain

Posted: 15 Jan 2022 09:25 AM PST

Let $i$ be a transient state of a continuous-time Markov chain $X$ with $X(0) = i$. Furthermore $X$ has right continious paths and the $Q$-matrix is stable and conservative. How do I show, that the total time spent in state $i$ has an exponential distribution and how do I find it´s rate? I suspect, that showing the memoryless property holds is sufficient, however I do not even know where to start. Any hints would be appreciated.

Study instability of a system with Lyapunov functions in terms of a parameter

Posted: 15 Jan 2022 09:42 AM PST

I am given the following system of odes \begin{gather} \begin{pmatrix} \dot{x}\\ \dot{y} \end{pmatrix} = \begin{pmatrix} 0 & 1\\ -1 & 0 \end{pmatrix} \begin{pmatrix} x\\ y \end{pmatrix} - b(x^2 + y^2) \begin{pmatrix} x\\ y \end{pmatrix} + (x^2 + y^2)^2 \begin{pmatrix} x\\ y \end{pmatrix} \end{gather} I want to study the stability of the $(0, 0)$ solution in terms of $b$. I cannot understand the argument they are using. It goes as follows: \begin{equation} V(x, y) = \frac{1}{2}(x^2+y^2) \end{equation} is a Lyapunov function for $(0, 0)$ under certain conditions. Exactly,

  1. $V(0, 0) = 0$
  2. $\dot{V}(x, y) = DV(x,y) f(x, y) = \\ = x(y-bx(x^2+y^2) + x(x^2+y^2)^2) + y(-x -by(x^2+y^2)+y(x^2+y^2)^2) = \\ = -b (x^2+y^2)^2 + (x^2+y^2)^3$

Then, if $b>0$ for all $(x,y)$ such that $x^2+y^2 < \sqrt{b}$ we have that $\dot{V} < 0$, therefore $V$ is a strict Lyapunov function and the solution $(0,0)$ is asymptotically stable.

If $b \leq 0$ then $\dot{V} > 0$ for all $(x, y) \neq (0, 0)$, therefore it is inestable (because if time goes the other way it would be asymptotically stable).

The parts I am not understanding are in bold.

First of all, when we talk about stability I thought solutions only needed to be defined for positive time. Therefore, by time "going the other way", what do we mean? Are we considering $-t$? Also, I don't see why this would mean instable.

Thanks in advance. Any help would be appreciated.

simple walk on $\mathbb{Z}$

Posted: 15 Jan 2022 09:20 AM PST

Consider the simple random walk $(X_n)_{n \in \mathbb{N}}$ starting from $X_0 = 0$.

Consider $\varepsilon>0$, show that, for all $\delta>0$, $$ \lim _{n \rightarrow \infty} \mathbb{P}\left(\frac{\left|X_{n}\right|}{n^{\frac{1}{2}+\varepsilon}}>\delta\right)=0 . $$

I'm new to random walks in probability and don't know how to solve this.

$\lim_{n \to \infty}\frac{\sum_{i=0}^{n}{\sqrt{i}}}{n^2}=0 \; ?$

Posted: 15 Jan 2022 09:49 AM PST

I was wondering whether this limit converges to zero: $$ \lim_{n \to \infty}\frac{\sum_{i=0}^{n}{\sqrt{i}}}{n^2}=0 $$ And i'm pretty sure it is.

First, by intuition. I know that $\sum_{i=0}^n{i} = \frac{n(n+1)}{2}$ ~ $O(n^2)$, so i guess that $\sum_{i=0}^n{\sqrt{i}}$ is "less powerful", but i don't really know how much "lesser"

So, the thing that really interest me was: what is the "cardinality" of $\sum_{i=0}^{\infty}{\sqrt{i}} \;? \quad$ Is it "equal ~" to $O(n)$? (I'm not sure i translated the words correctly. Does 'cardinality' is the right word for my description? I'm not familiar with these words in english, sorry. hope you understood what i meant).

Here is my thought:

$ {\sum_{i=0}^{n}{\sqrt{i}}} = {\sqrt{1}+\sqrt{2}+\sqrt{3}+...+\sqrt{n-1}+\sqrt{n}} = {\sqrt{n} \big( \frac{\sqrt{1}}{\sqrt{n}} +\frac{\sqrt{2}}{\sqrt{2}} +...+ \frac{\sqrt{n-1}}{\sqrt{n}} + \frac{\sqrt{n}}{\sqrt{n}} \big)} = {{\sqrt{n} \big( \sqrt{\frac{1}{n}} + \sqrt{\frac{2}{n}} +...+ \sqrt{\frac{n-1}{n}} + \sqrt{\frac{n}{n}} \big)}} $

and so:

$$ \lim_{n \to \infty} \frac{\sum_{i=0}^{n}{\sqrt{i}}}{n^2} = \lim_{n \to \infty} \frac{{\sqrt{n} \big( \sqrt{\frac{1}{n}} + \sqrt{\frac{2}{n}} +...+ \sqrt{\frac{n-1}{n}} + \sqrt{\frac{n}{n}} \big)}}{n^2} \overbrace{<}^{(*)} \lim_{n \to \infty} \frac{{\sqrt{n} \big( \sqrt{\frac{n}{n}} + \sqrt{\frac{n}{n}} +...+ \sqrt{\frac{n}{n}} + \sqrt{\frac{n}{n}} \big)}}{n^2} = \lim_{n \to \infty} \frac{{\sqrt{n} \big( \overbrace{1 + 1 +...+ 1 + 1}^{n \, times} \big)}}{n^2} = \lim_{n \to \infty} \frac{ \sqrt{n} *n }{n^2} = \lim_{n \to \infty} \frac{ \sqrt{n} }{n} = \lim_{n \to \infty} \frac{1}{ \sqrt{n} } = 0 $$

But my enlargement in $(*)$ above was too big.

I was wondering if you can suggest me some better way :)

Volume of unit nuclear norm ball

Posted: 15 Jan 2022 09:49 AM PST

Let $M_{n\times m}$ denote the vector space of $n \times m$ matrices. The nuclear norm is defined by

$$ \| M \|_* := \sum_{i=1}^{\min(m,n)} \sigma_i (M) $$

where $\sigma_i (M)$ are the singular values of $M$. The unit ball is defined as

$$ B^*_{nm}(1)=\{M\in M_{n\times m}: \|M\|_* \leq 1\} $$

What is the volume of $B^{*}_{nm}(1)$? What is the surface area?

The motivation for this question is just simple curiosity.

Exercise 13, Section 17 of Munkres’ Topology And Using This Result to Prove $\{x|f(x)=g(x)\}$ is Closed in $X$ [duplicate]

Posted: 15 Jan 2022 09:22 AM PST

Show that $X$ is Hausdorff if and only if the diagonal $\Delta =\{ x\times x|x\in X\}$ is closed in $X\times X$.

My attempt: first suppose $X$ is Hausdorff. let $(X\times X)-\Delta=\{ m\times n\in X\times X| m\neq n\}$. Let $x,y\in X$ with $x\neq y$, $\exists U_x, U_y\in \mathcal{T}_{X}$ such that $U_x\cap U_y=\phi$, by Hausdorff property of $X$. $\mathcal{B}=\{ U\times V|U,V\in \mathcal{T}_X\}$ is basis for $\mathcal{T}_p$, product topology. We want to show $(X\times X)-\Delta \in \mathcal{T}_p$. Claim: $(X\times X)-\Delta= \bigcup_{x\times y\in (X\times X)-\Delta} (U_x \times U_y)$. For $x\times y\in (X\times X)-\Delta, \exists U_x \times U_y\in \mathcal{B}$(existence depends on Hausdorff property) such that $x\times y\in U_x \times U_y \subseteq \bigcup_{x\times y\in (X\times X)-\Delta} (U_x \times U_y)$. Thus $(X\times X)-\Delta \subseteq \bigcup_{x\times y\in (X\times X)-\Delta} (U_x \times U_y)$. Now let $p\times q\in \bigcup_{x\times y\in (X\times X)-\Delta} (U_x \times U_y)$. Then $p\times q\in U_x \times U_y$, for some $x\times y\in (X\times X)-\Delta$. Since $U_x \cap U_y=\phi$ i.e. $\nexists z\in X$ such that $z\in U_x \cap U_y$. So $U_x \times U_y=\{ r\times s|r\in U_x, s \in U_y\}\subseteq (X\times X)-\Delta$, since $r\neq s$ (by $U_x \cap U_y=\phi$). Thus $(X\times X)-\Delta \supseteq \bigcup_{x\times y\in (X\times X)-\Delta} (U_x \times U_y)$. Hence $\Delta$ is closed in $X\times X$.

Conversely, suppose $\Delta$ is closed in $X\times X$. So $(X\times X)-\Delta=\{m\times n\in X\times X|m\neq n\} \in \mathcal{T}_p$. $(X\times X)-\Delta=\bigcup_{i\in I}(U_i\times V_i)$, where $(U_i\times V_i) \in \mathcal{B}$. Let $x\times y\in (X\times X)-\Delta$. Then $x\times y\in U_j \times V_j$, for some $j\in I$. Claim: $U_j\cap V_j=\phi$. Assume towards contradiction i.e. $U_j \cap V_j \neq \phi$. That means $\exists p\in U_j \cap V_j$. $p\in U_j$ and $p\in V_j$. So $p\times p\in U_j \times V_j$. Which implies $p\times p\in \bigcup_{i\in I}(U_i\times V_i)$. Thus $p\times p\in (X\times X)-\Delta$. Which contradicts the definition of $(X\times X)-\Delta$. Our initial assumption is wrong. Hence $U_j\cap V_j =\phi$.


lemma: Let $f, g : X \to Y$ be continuous; assume $Y$ is Hausdorff. Show that $\{x \mid f(x) = g(x)\}$ is closed in $X$.

You may have thought(I did) to prove this lemma by somehow showing inverse image of a closed set of continuous map is precisely $\{x|f(x)=g(x)\}$ set, hence closed in $X$. But constructing that map is not obvious/natural, at first glance(at least in my case). So later, I drop this strategy to prove above property of Hausdorff space. In my opinion, this is a candidate approach, one should consider.

Proof: https://math.stackexchange.com/a/3177710/861687

If $f(x)=\frac{1}{3} \biggl ( f(x+1)+\frac{5}{f(x+2)}\biggl)$ and $f(x)>0$, $\forall$ $x\in \mathbb R$ then $\lim_{x \rightarrow \infty} f(x)$ is?

Posted: 15 Jan 2022 09:51 AM PST

If $f(x)=\frac{1}{3} \biggl ( f(x+1)+\frac{5}{f(x+2)}\biggl)$ and $f(x)>0$, $\forall$ $x\in \mathbb R$ then $\lim_{x \rightarrow \infty} f(x)$ is?

Doubt:

In solution provided in book they assumed that

$\displaystyle \lim_{x \rightarrow \infty} f(x)=\displaystyle \lim_{x \rightarrow \infty}f(x+1)=\displaystyle \lim_{x \rightarrow \infty}f(x+2)=l$.

How can they all be equal?

For a square matrix $X$, do we have that: $\|X\|_2^2=\|X^TX\|_2$?

Posted: 15 Jan 2022 09:48 AM PST

Let $X$ be a $n\times n$ matrix. I think that $\|X\|_2^2 = \|X^{\top}X\|_2$, because:

$$\|X^{\top}X\|_2^{2} = \max_{\|x\|_{2}=1} x^{\top}\left(X^{\top}X\right)^2x=\rho\left(\left(X^{\top}X\right)^2\right)=\left(\rho\left(X^{\top}X\right)\right)^2 = \|X\|_{2}^{4}$$

Where $\rho$ denotes the maximum eigenvalue / spectral radius of $X$.

Is this calculation correct?

Secret sharing scheme, trying to figure out the cipher representation in the case of a polynomial function of degree $t$

Posted: 15 Jan 2022 09:32 AM PST

Suppose that we want to share a secret $s\in S$ in $t$ parts and we use a secret sharing scheme to do it. Suppose that $|Y|\geq |S|$ is a field sufficiently large with cardinality $p$ and a polynomial of degree $t$ where $(a_i)_{i=1}^t\in Y$ such that

$$f(x)=a_tx^t+a_{t_1}x^{t-1}+a_1x+a_0=s+\sum_{i=1}^ta_ix^i$$

such that $a_0=s$. In order to calculate the specific polynomial we want $t+1$ points say $(x_i,y_i)$ that determine uniquely the polynomial $f$ with constant $s$. In other words this relation that $t+1$ - pairs associated with only one $s$ is the cipher, say $c$. If we were in the case of three players a linear $f$ and a pair of numbers $(a,b)$ would be enough to determine $f(x)=a_1x+a_0$. So in the latter case we could say that the cipher $c(s,b)=a$ is bijective and the pair $(a,b)$ is associated with only one $s$ such that $$a+b\bmod{p}=s,\quad\text{$(1)$}$$

How is the cipher is going to be written in the case of $t>1$? Namely, if $(a_1,b_1),\dots,(a_{t+1}b_{t+1})$ are $t+1$ pairs from the polynomial $h$ of degree $t$ what is $c(s,b_1,b_2,\dots,b_{t+1})$ equals with and how the module equation is going to change in $(1)$?

What does it mean for the sequence $\text{Hom}(E,\text{Hom}(N,P))$ to be exact, where $E$ is an exact sequences of $A$-modules, and $N,P$ $A$-modules?

Posted: 15 Jan 2022 09:41 AM PST

Here $E$ is the exact sequence $M^{\prime} \stackrel{f}{\rightarrow} M \stackrel{g}{\rightarrow} M^{\prime \prime} \rightarrow 0$. I know what an exact sequence of $A$-modules is, but not what $\text{Hom}(E,\text{Hom}(N,P))$ means and what it means for it to be exact.

Does $y = f(x)$ and $y=g(x)$ imply $f(x)=g(x)$?

Posted: 15 Jan 2022 09:50 AM PST

Okay, this maybe a very dumb question but I can't seem to find a "proper" reason to show why this isn't true.

In some calculus books or notes, I have come across questions where they sometimes begin by saying, "If $y =f(x)$ and $y =g(x)$ are two functions and blah-blah..." I'm curious, is this some "abuse of notation" thing because it should be clear from the context? Because should it not mean that $y =f(x) = g(x)$?.

There is also another context where this is used— While making graphs. The question usually says graph $y=f(x)$ and $y=g(x)$ etc.

Surely it is not a substitution, otherwise the two functions would be same. So what does equating (possibly multiple) functions to $y$ mean? Or what is the intended meaning when we say $y =f(x)$ and $y =g(x)$?

Prove the associative law [closed]

Posted: 15 Jan 2022 09:46 AM PST

A is a set with operation $*$ satisfy two conditions

  1. $x*x=x$
  2. $(x*y)*z=(y* z)*x$

Prove $*$ is associative law and commutative is there idea about prove that assocaitve law first.

Calculating Gabriel's Trumpet volume $\left(\frac{1}{z}\right)$

Posted: 15 Jan 2022 09:45 AM PST

I want to calculate the volume of the solid obtained by rotating the function $1/z$ about the $z$-axis for $z>1$, that is $\pi$. But I want to apply that: \begin{equation*} V = 4\int \int_{\mathcal{S}} f(x,y) \,\text{d}x\text{d}y, \end{equation*} where \begin{equation*} z = f(x,y) = \frac{1}{\sqrt{x^2+y^2}}, \end{equation*} and the region $\mathcal{S}$ is $x^2+y^2\leq1$ for the first quadrant.

We change $x=r\cos\theta$ and $y=r\sin\theta$. Then \begin{equation*} V = 4\int \int_{\mathcal{S}} f(x,y) \,\text{d}x\text{d}y = 4\int \int_{\mathcal{T}} f(r,\theta) \frac{\partial(x,y)}{\partial(r,\theta)}\,\text{d}r\text{d}\theta. \end{equation*} Since $f(r,\theta)=1/r$ and $\frac{\partial(x,y)}{\partial(r,\theta)}=r$ we obtain \begin{equation*} V = 4\int_{0}^{\pi/2} \text{d}\theta \int_{0}^{1} \text{d}r = 2\pi \neq \pi \end{equation*}

Why I can't apply that method? Thanks.

Find the area of ​the shaded region in the figure below

Posted: 15 Jan 2022 09:22 AM PST

For reference:

In the figure $O$ is the center of the circle and its radius measures $a$ and $AQ = QB$. Calculate the area of ​​the shaded region.(Answer: $\frac{a^2}{4}(\pi-2)$)

enter image description here

My progress:

If $AQ = BQ \implies \angle AQB=90^\circ$

Complete the square $AQBD$.

image2

$OC = r$ and $QC =R = AC.$

$O$ is centre of square.

$QO$ is angle bisector, therefore $\angle AQO$ is $45^\circ.$

$QD = R\sqrt2$

Considering $\triangle OCQ$,

$\displaystyle r^2+\left(\frac{R}{2}\right)^2=OQ^2\implies r^2+\frac{R^2}{4}=(R\sqrt2)^2$

$\therefore R = \dfrac{2r\sqrt7}{7}$

I don't see a solution...is it missing some information?

Deduce the following set is not empty [closed]

Posted: 15 Jan 2022 09:23 AM PST

Source:Conover,A First Course In Topology

Problem 4.5.3.d (pg 147)

(This exercise requires knowledge of ordinals) We are going to show a function f on $[0,\Omega)$ is ultimately constant.This outline is based on Chapter 5 of Gillman and Jerison [6]. Conover gives the warning this is not easy.

I did a lot of googling and searching on MSE to solve them. They are $4$ parts:

a) If $A$ and $B$ are closed subsets of $[0,\Omega)$ then at least $A$ or $B$ is bounded. Proved.

b) show that for each ordinal number $a \in [0,\Omega)$, $[0,\Omega)$ is limit compact. Proved.

c) show a continuous image of a limit compact space is limit compact. Proved.

d.i) Let $f:[0,\Omega)\to R$ be continuous and show that for each ordinal number $a\in [0,\Omega )$, $f(0,\Omega))$ is a compact set. Proved.

I am stuck on this part.

Claim

ii)Deduce that $\bigcap \{f(a,\Omega)):a\in [0,\Omega)\}\ne\emptyset$ in $ R$.

Best guess

Since $f(a,\Omega))$ is a compact set, it has a cover with a finite subcover in it. So it is not empty? If my guess is correct I have to give that cover and show it so.

As for the last part

Let $r$ be a real number. So $r\in \bigcap \{f(a,\Omega)):a\in [0 ,\Omega )\}$. Then $$\bigcap\{f(a,\Omega)):0<a<r<\Omega\}.$$

How to treat an isomorphism between vector spaces as an isomorphism between many sorted structures

Posted: 15 Jan 2022 09:38 AM PST

According to this PDF, p.29, a vector space is a strict many sorted structure, where strict means that the scalar set and the vector set are disjoint (the definition is here, p.23).

For any field $K$, however, $K$ is also a vector space over $K$ and the scalar set and the vector set are the same, and therefore not disjoint. So we have to adopt a lax structure, where lax means not-strict(the definition is here, p.23), to formalize this vector space as many sorted structure.

But if we adopt a lax structure, (it seems to me) another problem arises.

Let $V$ be a one-dimensional vector space over $K$ satisfying $V \cap K = \emptyset$. It is well known that $V$ is isomorphic to the vector space $K$ mentioned above. But we cannot create the isomorphism between $V$ and $K$ as an isomorphism between many sorted structures(the definition is here, p.40), because if $\pi: K \cup V \to K \cup K = K$ is a isomorphism, $\pi\upharpoonright K: K \to K$ must be a bijection, but this is impossible since $\pi$ is a bijection between $K \cup V$ and $K$.

So I think we must modify or extend the definition of an isomorphism between many sorted structure to treat an isomorphism between vector spaces as an isomorphism between many sorted structures.

I would appreciate it if you could show me how to do this.

Why the space of section of a vector bundle is complete?

Posted: 15 Jan 2022 09:35 AM PST

This page is about the space of sections:

Let $E \stackrel{\mathrm{fb}}{\rightarrow} \Sigma$ be a smooth vector bundle. On its real vector space $\Gamma_{\Sigma}(E)$ of smooth sections consider the seminorms indexed by a compact subset $K \subset \Sigma$ and a natural number $N \in \mathbb{N}$ and given by $$ \begin{aligned} \Gamma_{\Sigma}(E) & \stackrel{p_{K}^{N}}{\longrightarrow} [0,\infty)\\ \Phi & \longmapsto \max _{n \leq N}\left(\sup _{x \in K}\left|\nabla^{n} \Phi(x)\right|\right) \end{aligned} $$ where on the right we have the absolute values of the covariant derivatives of $\Phi$ for any fixed choice of connection on $E$ and norm on the tensor product of vector bundles $\left(T^{*} \Sigma\right)^{\otimes_{\Sigma}^{n}} \otimes_{\Sigma} E$.

How to prove that this space is complete in respect to the seminorm?

Now In this article on page 6 GREEN-HYPERBOLIC OPERATORS ON GLOBALLY HYPERBOLIC SPACETIMES the author says that an Arzelà-Ascoli argument would show this but I don't see how.

If anyone would like me to expand the question please let me know.

When an ordinary partial derivative is a Frechet derivative on a Banach space?

Posted: 15 Jan 2022 09:49 AM PST

We have a function $p(x, \theta)\in \mathbb{R},\ x\in\mathbb{R},\ \theta\in\mathbb{R}^n$.

We can also consider $p(\cdot, \theta)$ to be a map from $\mathbb{R}^n$ to a Banach space of functions. For example, assuming $p(\cdot, \theta)\in \mathbb{L}^1(\mathbb{R})$ for each $\theta$ we have that $\hat p:\theta\to p(\cdot, \theta)$ is a map from $\mathbb{R}^n$ to $\mathbb{L}^1(\mathbb{R})$.

My question is about relation of the ordinary partial derivative $\frac{\partial p(x, \theta)}{\partial\theta}$ (assuming it is $\in\mathbb{L}^1(\mathbb{R})$) to the Frechet derivative of $\hat p:\theta\to p(\cdot, \theta)$. When are they the same?

Addition after the comments of Zerox:

For the ordinary partial derivative for each $x$ $$ \lim_{d\theta\to0}\frac{\left|p(x,\theta+d\theta)-p(x, \theta)-\frac{\partial p(x, \theta)}{\partial\theta}\cdot d\theta\right|}{\Vert d\theta \Vert}=0 $$ For the Frechet derivative we have different convergence $$\lim_{d\theta\to0}\frac{\Vert p(\cdot,\theta+d\theta)-p(\cdot, \theta)-L(d\theta)\Vert_{\mathbb{L}^1}}{\Vert d\theta\Vert}=0$$ where $L(d\theta)$ is a continuous linear map from $\mathbb{R}^n$ to $\mathbb{L}^1(\mathbb{R})$.

The linear map $d\theta\to\frac{\partial p(\cdot, \theta)}{\partial\theta}\cdot d\theta$ from $\mathbb{R}^n$ to $\mathbb{L}^1(\mathbb{R})$ is continuous in the Banach norm. $$\left\|\frac{\partial p(\cdot, \theta)}{\partial\theta}\cdot d\theta\right\|_{\mathbb{L}^1}=\left\|\sum_i\frac{\partial p(\cdot, \theta)}{\partial\theta_i}d\theta_i\right\|_{\mathbb{L}^1}\leq\sum_i\left\Vert\frac{\partial p(\cdot, \theta)}{\partial\theta_i}\right\Vert_{\mathbb{L}^1}|d\theta_i|\leq\operatorname{max}_i\left\Vert\frac{\partial p(\cdot, \theta)}{\partial\theta_i}\right\|_{\mathbb{L}^1}\Vert d\theta\Vert_{\mathbb{L}^1}$$

So this map seems to be a candidate for $L(d\theta)$, but i don't see why (or when) they have to coincide. Pointwise convergence (in $x$) doesn't imply $\mathbb{L}^1$ convergence and vice versa.

Of course in the topology of pointwise convergence on the Banach space they would coincide, but i am interested in the $\mathbb{L}^1$ convergence.

A special case of parametrized probability density functions:

If there are no convenient conditions for the general case, then maybe there are some for this case? $$\int p(x,\theta)dx=1,\ p(x,\theta)\geq0\ \forall\theta$$

Upd: deleted the incorrect Scheffe's Lemma application.

The case of dominated partial derivatives:

Assume $\left|\frac{\partial p(x, \theta)}{\partial\theta_i}\right|\leq g(x)$ for some $g(x)\in \mathbb{L}^1(\mathbb{R})$. Then by the multivariate MVT $$ \begin{split} \frac{\left|p(x,\theta+d\theta)-p(x, \theta)-\frac{\partial p(x, \theta)}{\partial\theta}\cdot d\theta\right|}{\Vert d\theta\Vert}\leq&\frac{\left|\frac{\partial p(x, c)}{\partial\theta}\cdot d\theta\right|+\left|\frac{\partial p(x, \theta)}{\partial\theta}\cdot d\theta\right|}{\Vert d\theta\Vert}\\ \leq&\left\Vert\frac{\partial p(x, c)}{\partial\theta}\right\Vert+\left\Vert\frac{\partial p(x, \theta)}{\partial\theta}\right\Vert\leq2\sqrt n g(x) \end{split} $$ So the whole expression is also dominated by the $\mathbb{L}^1(\mathbb{R})$ function.

By the dominated convergence theorem we have the convergence of the expression above in $\mathbb{L}^1$, so $L(d\theta)=\frac{\partial p(x, \theta)}{\partial\theta}\cdot d\theta$.

If both derivatives exist, they have to be equal:

One more observation. For any $\mathbb R \ni h_m\to0$ the existence of the Frechet derivative implies the $\mathbb L^1$ convergence to $0$ of the corresponding expression for each $i=1,\cdots,n$. $$ \left\Vert\frac{p(\cdot,\theta+e_ih_m)-p(\cdot, \theta)}{h_m}-L(e_i)\right\Vert_{\mathbb{L}^1}\to0$$ Where $e_i$ is $i$th vector of the standard basis.

Then there exists a subsequence of $h_m$, call it $\mathbb R \ni g_l\to0$, for which the expression above converges pointwise a.e. to $0$. $$\left|\frac{p(x,\theta+e_ig_l)-p(x, \theta)}{g_l}-L(e_i)(x)\right|\to0\ \textbf{a.e.}$$ But we also have for the ordinary partial derivative $$\left|\frac{p(x,\theta+e_ig_l)-p(x, \theta)}{g_l}-\frac{\partial p(x, \theta)}{\partial\theta_i}\right|\to0$$

So $L(e_i)=\frac{\partial p(\cdot, \theta)}{\partial\theta_i}$and thus $L(d\theta)=\frac{\partial p(\cdot, \theta)}{\partial\theta}\cdot d\theta$.

Additions, corrections, indications of errors are still welcome.

How do you prove $S_{XYZT} \leq \dfrac{1}{5} S_{ABCD} $?

Posted: 15 Jan 2022 09:30 AM PST

In the given figure, $ABCD$ is a convex quadrilateral. Suppose that $M, N, P, Q$ are mid-points of $AB, BC, CD, DA$, respectively. Prove that $S_{XYZT} \leq \dfrac{1}{5} S_{ABCD} $ where $S_{ABCD}$ (resp. $S_{XYZT}$) is the area of $ABCD$ (resp. $XYZT$)?

enter image description here

Could you please give a key hint to solve this exercise? Thank you so much for your discussions!

Collatz conjecture, Tao-Collatz remainder and mod n.

Posted: 15 Jan 2022 09:22 AM PST

Collatz conjecture is equivalent to $n\times 3^{k} = 2^{ak+1} - TCR$ where, for me, $k$=odd steps, and $ak+1 $=even steps. Note that total steps = k +( ak+1) steps. Some numbers have the same total steps, but $k$ steps and $ak+1$ steps are different because it is unique for every n.

Even numbers are $\equiv \pmod{2}$ remainder 0 and $\equiv\pmod{3}$ remainder 0,1,2 . Odd numbers are $ \equiv \pmod{2}$ remainder 1 and $\equiv\pmod{3}$ remainder 0,1,2.

Professor Terence Tao talks about $3^{k-1}*2^{a1} + 3^{k-2}*2^{a2} + ... +2^{ak }$ and the study of his properties to take any advance in the solution on his post about Collatz conjecture. Let´s call it Tao-Collatz remainder, $TCR$ for me, I think that

$2^{ak+1}\equiv TCR \pmod{n}$

The gcd ($2^{ak+1}$, n) =1 , so there exists just 1 solution for this congruence.And $2^{ak+1}$ – TCR is a multiple of $n$. Effectively for each power of 2, $2^{x}> n\times 3^{k}$ exist 1 solution with the form $2^{x}-r$ .This simple question based on the Euclidean algorithm of division may not be strong enought for a demostration. We know that

$n^{i}\equiv r_{i}\pmod{n}$

When $3^{0}\rightarrow TCR= 0$ , $3^{1}\rightarrow TCR= 1$ , $3^{k}\geq 2\rightarrow TCR = 3^{k-1}*2^{a1} + 3^{k-2}*2^{a2} + ... +2^{ak }$ And

$3^{0}\equiv 1\pmod{n}$ , $3^{1}\equiv 3\pmod{n}$ , $k\geq 2 ,3^{k}\equiv \chi \pmod{n}$

If $2^{ak+1}\chi \equiv TCR \pmod{n}$ and $3^{k-1}\equiv \chi \pmod{n}$, where $\chi $ is the remainder, is the solution, and the Euclidean algorithm always ends, ¿is it the probe that the Collatz conjecture is true?

What is the solution to this congruence?

$n_1,...,n_k$ pair coprime $\!\iff\! {\rm lcm}(n_1,...,n_k)=n_1...n_k$ [lcm = product for coprimes] [closed]

Posted: 15 Jan 2022 09:45 AM PST

$n_1,...,n_k$ pairwise coprime $\iff LCM(n_1,...,n_k)=n_1...n_k$

Recently, I was told this as part of a larger proof concerning direct products of groups. I am wondering why this is true.

Continuous image of the intersection of decreasing sets in a compact space

Posted: 15 Jan 2022 09:50 AM PST

Suppose $B_{\epsilon}$ are closed subsets of a compact space and $B_{\epsilon} \supset B_{\epsilon'} \quad \forall \epsilon > \epsilon'$. Furthermore, $B_0 = \bigcap_{\epsilon>0} B_{\epsilon}$. For a continuous function $f$ can we conclude that $$f(B_0) = \bigcap_{\epsilon>0} f(B_{\epsilon})?$$

I believe the answer to be yes. It seems this should be a well-known property---I'm having trouble finding a reference.

No comments:

Post a Comment