$\ell^p$ and $L^p$ as Metric Spaces

Let $p\geq 1$ be a fixed number and let
$$\ell^p=\left\{x=(\xi_j): \sum_{j=1}^\infty|\xi_j|^p<\infty\right\}.$$
Define $d:\ell^p\times\ell^p\longrightarrow\mathbb{R}^+\cup\{0\}$ by
$$d(x,y)=\left(\sum_{j=1}^\infty|\xi_j-\eta_j|^p\right)^{\frac{1}{p}}.$$
Then $(\ell^p,d)$ is a metric space. The properties (M1) and (M2) are clearly satisfied. We prove the remaining property (M3) the triangle inequality. $p=1$ case can be easily shown by the triangle inequality of numbers. We need a few steps to do this. First we prove the following inequality: $\forall\alpha>0,\beta>0$,
$$\alpha\beta\leq\frac{\alpha^p}{p}+\frac{\beta^q}{q},$$
where $p>1$ and $\frac{1}{p}+\frac{1}{q}=1$. The numbers $p$ and $q$ are called conjugate exponents. It follows from $\frac{1}{p}+\frac{1}{q}=1$ that $(p-1)(q-1)=1$ i.e. $\frac{1}{p-1}=q-1$. If we let $u=t^{p-1}$ then $t=u^{\frac{1}{p-1}}=u^{q-1}$. By comparing areas, we obtain
$$\alpha\beta\leq\int_0^{\alpha}t^{p-1}dt+\int_0^{\beta}u^{q-1}du=\frac{\alpha^p}{p}+\frac{\beta^q}{q}.$$
Next, using this inequality we prove the Hölder inequality
$$\sum_{j=1}^\infty|\xi_j\eta_j|\leq\left(\sum_{k=1}^\infty|\xi_k|^p\right)^{\frac{1}{p}}\left(\sum_{m=1}^\infty|\eta_m|^q\right)^{\frac{1}{q}}$$
where $p>1$ and $\frac{1}{p}+\frac{1}{q}=1$. When $p=2$ and $q=2$, we obtain the well-known Cauchy-Schwarz inequality.

Proof. Let $(\tilde\xi_j)$ and $(\tilde\eta_j)$ be two sequences such that
$$\sum_{j=1}^\infty|\tilde\xi_j|^p=1,\ \sum_{j=1}^\infty|\tilde\eta_j|^q=1.$$
Let $\alpha=|\tilde\xi_j|$ and $\beta=|\tilde\eta_j|$. Then by the inequality we proved previously,
$$|\tilde\xi_j\tilde\eta_j|\leq\frac{|\tilde\xi_j|^p}{p}+\frac{|\tilde\eta_j|^q}{q}$$
and so we obtain
$$\sum_{j=1}^\infty|\tilde\xi_j\tilde\eta_j|\leq\sum_{j=1}^\infty\frac{|\tilde\xi_j|^p}{p}+\sum_{j=1}^\infty\frac{|\tilde\eta_j|^q}{q}=1.$$
Now take any nonzero $x=(\xi_j)\in\ell^p$, $y=(\eta_j)\in\ell^q$. Setting
$$\tilde\xi_j=\frac{\xi_j}{\left(\displaystyle\sum_{k=1}^\infty|\xi_k|^p\right)^{\frac{1}{p}}},\ \tilde\eta_j=\frac{\eta_j}{\left(\displaystyle\sum_{m=1}^\infty|\eta_m|^q\right)^{\frac{1}{q}}}$$
results in the Hölder inequality.

Next, we prove the Minkowski inequality
$$\left(\sum_{j=1}^\infty|\xi_j+\eta_j|^p\right)^{\frac{1}{p}}\leq\left(\sum_{k=1}^\infty|\xi_k|^p\right)^{\frac{1}{p}}+\left(\sum_{m=1}^\infty|\eta_m|^p\right)^{\frac{1}{p}}$$
where $x=(\xi_j)\,y=(\eta_j)\in\ell^p$ and $p\geq 1$. $p=1$ case comes from the triangle inequality for numbers. Let $p>1$. Then
\begin{align*}
|\xi_j+\eta_j|^p&=|\xi_j+\eta_j||\xi_j+\eta_j|^{p-1}\\
&\leq(|\xi_j|+|\eta_j|)|\xi_j+\eta_j|^{p-1}\ (\mbox{triangle inequality for numbers}).
\end{align*}
For a fixed $n$, we have
$$\sum_{j=1}^n|\xi_j+\eta_j|^p\leq\sum_{j=1}^n|\xi_j||\xi_j+\eta_j|^{p-1}+\sum_{j=1}^n|\eta_j||\xi_j+\eta_j|^{p-1}.$$
Using the Hölder inequality, we get the following inequality
\begin{align*}
\sum_{j=1}^n|\xi_j||\xi_j+\eta_j|^{p-1}&\leq \sum_{j=1}^\infty |\xi_j||\xi_j+\eta_j|^{p-1}\\
&\leq\left(\sum_{k=1}^\infty |\xi_k|^p\right)^{\frac{1}{p}}\left(\sum_{m=1}^\infty(|\xi_m+\eta_m|^{p-1})^q\right)^{\frac{1}{q}}\ (\mbox{Hölder})\\
&=\left(\sum_{k=1}^\infty|\xi_k|^p\right)^{\frac{1}{p}}\left(\sum_{m=1}^\infty|\xi_m+\eta_m|^p\right)^{\frac{1}{q}}.
\end{align*}
Similarly, we also get the inequality
$$\sum_{j=1}^n|\eta_j||\xi_j+\eta_j|^{p-1}\leq \left(\sum_{k=1}^\infty|\eta_k|^p\right)^{\frac{1}{p}}\left(\sum_{m=1}^\infty|\xi_m+\eta_m|^p\right)^{\frac{1}{q}}.$$
Combining these two inequalities, we get
$$\sum_{j=1}^n|\xi_j+\eta_j|^p\leq\left\{\left(\sum_{k=1}^\infty|\xi_k|^p\right)^{\frac{1}{p}}+\left(\sum_{k=1}^\infty|\eta_k|^p\right)^{\frac{1}{p}}\right\}\left(\sum_{m=1}^\infty|\xi_m+\eta_m|^p\right)^{\frac{1}{q}}$$
and by taking the limit $n\to \infty$ on the left hand side, we get
$$\sum_{j=1}^\infty|\xi_j+\eta_j|^p\leq\left\{\left(\sum_{k=1}^\infty|\xi_k|^p\right)^{\frac{1}{p}}+\left(\sum_{k=1}^\infty|\eta_k|^p\right)^{\frac{1}{p}}\right\}\left(\sum_{m=1}^\infty|\xi_m+\eta_m|^p\right)^{\frac{1}{q}}.$$
Finally, dividing this inequality by $\displaystyle\left(\sum_{m=1}^\infty|\xi_m+\eta_m|^p\right)^{\frac{1}{q}}$ results in the Minkowski inequality. The Minkowski inequality tells that
$$d(x,y)=\left(\sum_{j=1}^\infty|\xi_j-\eta_j|^p\right)^{\frac{1}{p}}<\infty$$
for $x,y\in\ell^p$. Let $x=(\xi_j), y=(\eta_j),\ z=(\zeta_j)\in\ell^p$. Then
\begin{align*}
d(x,y)&=\left(\sum_{j=1}^\infty|\xi_j-\eta_j|^p\right)^{\frac{1}{p}}\\
&\leq\left(\sum_{j=1}^\infty[|\xi_j-\zeta_j|+|\zeta_j-\eta_j|]^p\right)^{\frac{1}{p}}\\
&\leq\left(\sum_{j=1}^\infty|\xi_j-\zeta_j|^p\right)^{\frac{1}{p}}+\left(\sum_{j=1}^\infty|\zeta_j-\eta_j|^p\right)^{\frac{1}{p}}\\
&=d(x,z)+d(z,y).
\end{align*}The inequality that is second to the last expression is obtained by Minkowski inequality.

A measurable function $f$ on a closed interval $[a,b]$ is said to belong to $L^p$ if $\int_a^b|f(t)|^p dt<\infty$. $L^p$ is a vector space. For functions $f,g\in L^p$, we define
$$d(f,g)=\left\{\int_a^b|f(t)-g(t)|^pdt\right\}^{\frac{1}{p}}.$$
Then clearly (M2) symmetry is satisfied and one can also prove that (M3) triangle inequality holds. However, (M1) is not satisfied since what we have is that if $d(f,g)=0$ then $f=g$ a.e. (almost everywhere) i.e. the set $\{t\in[a,b]: f(t)\ne g(t)\}$ has measure $0$. It turns out that $=$ a.e. is an equivalence relation on $L^p$, so by considering $f\in L^p$ as its equivalence class $[f]$, $d$ can be defined as a metric on $L^p$ (actually the quotient space of $L^p$). Later, we will be particularly interested in the case when $p=2$ in which case $L^p$ as well as $\ell^p$ become Hilbert spaces. Those of you who want to know details about $L^p$ space are referred to

Real Analysis, H. L. Royden, 3rd Edition. Macmillan Publishing Company, 1988

Group Theory 2: Preliminaries (Functions)

In my previous notes here, I mentioned some about logical symbols. The logical symbols I will use often are $\forall$ which means “for all”, “for any”, “for each”, or “for every” depending on the context, $\exists$ which means “there exists”, and $\ni$ which means “such that” (don’t be confused with $\in$ which means “be an element of”). We also use s.t. for “such that.” There are also $\Longrightarrow$ which means “implies” and $\Longleftrightarrow$ which means “if and only if.” I guess these pretty much cover what we use most of time.

Now lets review about functions in a more formal way. Let $X$ and $Y$ be two non-empty sets. The the Cartesian product $X\times Y$ of $X$ and $Y$ is defined as the set
$$X\times Y=\{(x,y): x\in X,\ y\in Y\}.$$
A subset $f$ of the Cartesian product $X\times Y$ (we write $f\subset X\times Y$) is called a graph from $X$ to $Y$. A graph $f\subset X\times Y$ is called a function from $X$ to $Y$ (we write $f: X\longrightarrow Y$) if whenever $(x,y_1),(x,y_2)\in f$, $y_1=y_2$. If $f: X\longrightarrow Y$ and $(x,y)\in f$, we also write $y=f(x)$. A function $f: X\longrightarrow Y$ is said to be one-to-one or injective if whenever $(x_1,y),(x_2,y)\in f$, $x_1=x_2$. This is equivalent to saying $f(x_1)=f(x_2)$ implies $x_1=x_2$. A function $f: X\longrightarrow Y$ is said to be onto or surjective if $\forall y\in Y$ $\exists x\in X$ s.t. $(x,y)\in f$. A function $f: X\longrightarrow Y$ is said to be one-to-one and onto (or bijective) if it is both one-to-one and onto (or both injective and surjective).

Let $f: X\longrightarrow Y$ and $g: Y\longrightarrow Z$ be two functions. Then the composition or the composite function $g\circ f: X\longrightarrow Z$ is defined by $g\circ f(x)=g(f(x))$ $\forall x\in X$. The function composition $\circ$ may be considered as an operation and it is associative.

Lemma. If $h: X\longleftrightarrow Y$, $g:Y\longleftrightarrow Z$ and $f:Z\longleftrightarrow W$, then $f\circ(g\circ h)=(f\circ g)\circ h$.

Note that $\circ$ is not commutative i.e. it is not necessarily true that $f\circ g=g\circ f$ even when both $f\circ g$ and $g\circ f$ are defined.

The following lemmas will be useful when we study group theory later.

Lemma. If both $f: X\longrightarrow Y$ and $g: Y\longrightarrow Z$ are one-to-one, so is $g\circ f: X\longrightarrow Z$.

Lemma. If both $f: X\longrightarrow Y$ and $g: Y\longrightarrow Z$ are onto, so is $g\circ f: X\longrightarrow Z$.

As an immediate consequence of combining these two lemmas, we obtain

Lemma. If both $f: X\longrightarrow Y$ and $g: Y\longrightarrow Z$ are bijective, so is $g\circ f: X\longrightarrow Z$.

If $f\subset X\times Y$, then the inverse graph $f^{-1}\subset Y\times X$ is defined by
$$f^{-1}=\{(y,x)\in Y\times X: (x,y)\in f\}.$$
If $f: X\longrightarrow Y$ is one-to-one and onto (bijective) then its inverse graph $f^{-1}$ is a function $f^{-1}: Y\longrightarrow X$. The inverse $f^{-1}$ is also one-to-one and onto.

Lemma. If $f: X\longrightarrow Y$ is a bijection, then $f\circ f^{-1}=\imath_Y$ and $f^{-1}\circ f=\imath_X$, where $\imath_X$ and $\imath_Y$ are the identity mappings of $X$ and $Y$, respectively.

Let $A(X)$ be the set of all one-to-one functions of $X$ onto $X$ itself. Then $(A(X),\circ)$ is a group. If $X$ is a finite set of $n$-elements (we may conveniently say $X=\{1,2,\cdots,n\})$, then $(A(X),\circ)$ is a finite group of order $n!$, called the symmetric group of degree $n$. The symmetric group of degree $n$ is denoted by $S_n$ and the elements of $S_n$ are called permutations.

Analytic Continuation

The function $f(z)=\displaystyle\frac{1}{1+z}$ has an isolated singularity at $z=-1$. It has the Maclaurin series representation

$$f(z)=\sum_{n=0}^\infty(-1)^nz^n$$
for $|z|<1$. The power series $f_1(z)=\displaystyle\sum_{n=0}^\infty(-1)^nz^n$ converges only on the open unit disk $D_1:\ |z|<1$. For instance, the series diverges at $z=\frac{3}{2}i$ i.e. $f_1\left(\frac{3}{2}i\right)$ is not defined. The first 25 partial sums of the series $f_1\left(\frac{3}{2}i\right)$ are listed below and they do not appear to be approaching somewhere.

S[1] = 1.
S[2] = 1. – 1.500000000 I
S[3] = -1.250000000 – 1.500000000 I
S[4] = -1.250000000 + 1.875000000 I
S[5] = 3.812500000 + 1.875000000 I
S[6] = 3.812500000 – 5.718750000 I
S[7] = -7.578125000 – 5.718750000 I
S[8] = -7.578125000 + 11.36718750 I
S[9] = 18.05078125 + 11.36718750 I
S[10] = 18.05078125 – 27.07617188 I
S[11] = -39.61425781 – 27.07617188 I
S[12] = -39.61425781 + 59.42138672 I
S[13] = 90.13208008 + 59.42138672 I
S[14] = 90.13208008 – 135.1981201 I
S[15] = -201.7971802 – 135.1981201 I
S[16] = -201.7971802 + 302.6957703 I
S[17] = 455.0436554 + 302.6957703 I
S[18] = 455.0436554 – 682.5654831 I
S[19] = -1022.848225 – 682.5654831 I
S[20] = -1022.848225 + 1534.272337 I
S[21] = 2302.408505 + 1534.272337 I
S[22] = 2302.408505 – 3453.612758 I
S[23] = -5179.419137 – 3453.612758 I
S[24] = -5179.419137 + 7769.128706 I
S[25] = 11654.69306 + 7769.128706 I

Also shown below are the graphics of partial sums of the series $f_1\left(\frac{3}{2}i\right)$.

The first 10 partial sums

The first 10 partial sums

The first 20 partial sums

The first 20 partial sums

The first 30 partial sums

The first 30 partial sums

Let us expand $f(z)=\displaystyle\frac{1}{1+z}$ at $z=i$. Then we obtain
\begin{align*}
f(z)&=\frac{1}{1+z}\\
&=\frac{1}{1+i}\cdot\frac{1}{1+\frac{z-i}{1+i}}\\
&=\sum_{n=0}^\infty (-1)^n\frac{(z-i)^n}{(1+i)^{n+1}}
\end{align*}
for $|z-i|<\sqrt{2}$. Let $f_2(z)=\displaystyle\sum_{n=0}^\infty (-1)^n\frac{(z-i)^n}{(1+i)^{n+1}}$. This series converges only on the open disk $D_2:\ |z-i|<\sqrt{2}$, in particular at $z=\frac{3}{2}i$ and $f_2\left(\frac{3}{2}i\right)=f\left(\frac{3}{2}i\right)=\frac{4}{13}-\frac{6}{13}i$. The first 25 partial sums of the series $f_2\left(\frac{3}{2}i\right)$ are listed below and it appears that they are approaching a number. In fact, they are approaching the complex number $f\left(\frac{3}{2}i\right)=\frac{4}{13}-\frac{6}{13}i$.

S[1] = 0.5000000000 – 0.5000000000 I
S[2] = 0.2500000000 – 0.5000000000 I
S[3] = 0.3125000000 – 0.4375000000 I
S[4] = 0.3125000000 – 0.4687500000 I
S[5] = 0.3046875000 – 0.4609375000 I
S[6] = 0.3085937500 – 0.4609375000 I
S[7] = 0.3076171875 – 0.4619140625 I
S[8] = 0.3076171875 – 0.4614257812 I
S[9] = 0.3077392578 – 0.4615478516 I
S[10] = 0.3076782227 – 0.4615478516 I
S[11] = 0.3076934814 – 0.4615325928 I
S[12] = 0.3076934814 – 0.4615402222 I
S[13] = 0.3076915741 – 0.4615383148 I
S[14] = 0.3076925278 – 0.4615383148 I
S[15] = 0.3076922894 – 0.4615385532 I
S[16] = 0.3076922894 – 0.4615384340 I
S[17] = 0.3076923192 – 0.4615384638 I
S[18] = 0.3076923043 – 0.4615384638 I
S[19] = 0.3076923080 – 0.4615384601 I
S[20] = 0.3076923080 – 0.4615384620 I
S[21] = 0.3076923075 – 0.4615384615 I
S[22] = 0.3076923077 – 0.4615384615 I
S[23] = 0.3076923077 – 0.4615384616 I
S[24] = 0.3076923077 – 0.4615384615 I
S[25] = 0.3076923077 – 0.4615384615 I

The following graphics shows that the real parts of the partial sums of the series $f_2\left(\frac{3}{2}i\right)$ are approaching $\frac{3}{14}$ (blue line).

The real parts of the first 25 partial sums

The real parts of the first 25 partial sums

The next graphics shows that the imaginary parts of the partial sums of the series $f_2\left(\frac{3}{2}i\right)$  are approaching $-\frac{6}{13}$ (blue line).

The imaginary parts of the first 25 partial sums

The imaginary parts of the first 25 partial sums

Also shown below is the graphics of the first 25 partial sums of the series $f_2\left(\frac{3}{2}i\right)$. They are approaching the complex number $f\left(\frac{3}{2}i\right)=\frac{4}{13}-\frac{6}{13}i$ (the intersection of horizontal and vertical blue lines).

The first 25 partial sums

The first 25 partial sums

Note that $f_1(z)=f_2(z)$ on $D_1\cap D_2$. Define $F(z)$ as

$$F(z)=\left\{\begin{array}{ccc}
f_1(z) & \mbox{if} & z\in D_1,\\
f_2(z) & \mbox{if} & z\in D_2.
\end{array}\right.$$

Analytic continuation

Analytic continuation

Then $F(z)$ is analytic in $D_1\cup D_2$. The function $F(z)$ is called the analytic continuation into $D_1\cup D_2$ of either $f_1$ or $f_2$, and $f_1$ and $f_2$ are called elements of $F$. The function $f_1(z)$ can be continued analytically to the punctured plane $\mathbb{C}\setminus\{-1\}$ and the function $f(z)=\frac{1}{1+z}$ is indeed the analytic continuation into $\mathbb{C}\setminus\{-1\}$ of $f_1$. In general, whenever analytic continuation exists it is unique.

Metric Spaces

This is the first of series of lecture notes I intend to write on Functional Analysis.

What is functional analysis? Functional analysis is an abstract branch of mathematics, especially of analysis, concerned with the study of vector spaces of functions. These vector spaces of functions arise naturally when we study linear differential equations as solutions of a linear differential equation form a vector space. Functional analytic methods and results are important in various fields of mathematics (for example, differential geometry, ergodic theory, integral geometry, noncommutative geometry, partial differential equations, probability, representation theory etc.) and its applications, in particular, in economics, finance, quantum mechanics, quantum field theory, and statistical physics. Topics in this introductory functional analysis course include metric spaces, Banach spaces, Hilbert spaces, bounded linear operators, the spectral theorem, and unbounded linear operators.

While functional analysis is a branch of analysis, due to its nature linear algebra is heavily used. So, it would be a good idea to brush up on linear algebra among other things you need to study functional analysis.

In functional analysis, we study analysis on an abstract space $X$ rather than the familiar $\mathbb{R}$ or $\mathbb{C}$. In order to consider fundamental notions in analysis such as limits and convergence, we need to have distance defined on $X$ so that we can speak of nearness or closeness. A distance on $X$ can be defined as a function, called a distance function or a metric, $d: X\times X\longrightarrow\mathbb{R}^+\cup\{0\}$ satisfying the following properties:

(M1) $d(x,y)=0$ if and only if $x=y$.

(M2) $d(x,y)=d(y,x)$ (Symmetry)

(M3) $d(x,y)\leq d(x,z)+d(z,y)$ (Triangle Inequality)

Here $\mathbb{R}^+$ denotes the set of all positive real numbers. You can easily see how mathematicians came up with this definition of a metric. (M1)-(M3) are the properties that the familiar distance on $\mathbb{R}$, $d(x,y)=|x-y|$ satisfies. The space $X$ with a metric $d$ is called a metric space and we usually write it as $(X,d)$.

Example. Let $x=(\xi_1,\cdots,\xi_n), y=(\eta_1,\cdots,\eta_n)\in\mathbb{R}^n$. Define
$$d(x,y)=\sqrt{(\xi_1-\eta_1)^2+\cdots+(\xi_n-\eta_n)^2}.$$
Then $d$ is a metric on $\mathbb{R}^n$ called the Euclidean metric.

This time, let $x=(\xi_1,\cdots,\xi_n), y=(\eta_1,\cdots,\eta_n)\in\mathbb{C}^n$ and define
$$d(x,y)=\sqrt{|\xi_1-\eta_1|^2+\cdots+|\xi_n-\eta_n|^2}.$$
Then $d$ is a metric on $\mathbb{C}^n$ called the Hermitian metric. Here $|\xi_i-\eta_i|^2=(\xi_i-\eta_i)\overline{(\xi_i-\eta_i)}$.

Of course these are pretty familiar examples. If there can be only these familiar examples, there would be no point of considering abstract space. In fact, the abstraction allows to discover other examples of metrics that are not so intuitive.

Example. Let $X$ be the set of all bounded sequences of complex numbers
$$X=\{(\xi_j): \xi_j\in\mathbb{C},\ j=1,\cdots\}.$$
For $x=(\xi_j), y=(\eta_j)\in X$, define
$$d(x,y)=\sup_{j\in\mathbb{N}}|\xi_j-\eta_j|.$$
Then $d$ is a metric on $X$. The metric space $(X,d)$ is denoted by $\ell^\infty$.

Example. Let $X$ be the set of continuous real-valued functions defined on the closed interval $[a,b]$. Let $x, y:[a,b]\longrightarrow\mathbb{R}$ be continuous and define
$$d(x,y)=\max_{t\in [a,b]}|x(t)-y(t)|.$$
Then $d$ is a metric on $X$. The metric space $(X.d)$ is denoted by $\mathcal{C}[a,b]$.

In a metric space $(X,d)$, nearness or closeness can be described by a neighbourhood called an $\epsilon$-ball ($\epsilon>0$) centered at $x\in X$
$$B(x,\epsilon)=\{y\in X: d(x,y)<\epsilon\}.$$
These $\epsilon$-balls form a base for a topology on $X$, called the topology on $X$ induced by the metric $d$.

Next time, we will discuss two more examples of metric spaces $\ell^p$ and $L^p$. These examples are particularly important in functional analysis as they become Banach spaces. In particular, they become Hilbert spaces when $p=2$.

Group Theory 1: An Overview

This is the first of a series of lecture notes on group theory I intend to write for undergraduate Modern Algebra I course I am teaching in the fall semester. Before we begin to discuss the subject, I would like to give an overview of what we study in group theory or more generally in algebra.

Algebra (as a subject) is the study of algebraic structures. So, what is an algebraic structure? An algebraic structure or an algebra in short $\underline{A}$ is a non-empty set $A$ with a binary operation $f$. $\underline{A}$ is usually written as the ordered pair
$$\underline{A}=(A,f).$$
A binary operation $f$ on a set $A$ is a function $f: A\times A\longrightarrow A$. An example of a binary operation is addition $+$ on the set of integers $\mathbb{Z}$. $+$ is a function $+:\mathbb{Z}\times\mathbb{Z}\longrightarrow\mathbb{Z}$ defined by $+(1,1)=2$, $+(1,2)=3$, and so on. We usually write $+(1,1)=2$ as $1+1=2$. In general, one may consider an $n$-ary operation $f:\prod_{i=1}^n A\longrightarrow A$, where $\prod_{i=1}^n A$ denotes the $n$-copies of $A$, $A\times A\times\cdots\times A$.

There are many different kinds of algebras. Let me mention some of algebras with a binary operation here. For starter, $(A,\cdot)$, a non-empty set $A$ with a binary operation $\cdot$ is called a groupoid. A groupoid $(A,\cdot)$ with associative law
$$(ab)c=a(bc)$$
for any $a,b,c\in A$ is callaed a semigroup. If the semigroup has an identity element $e\in A$ i.e.
$$ae=ea=a$$
for any $a\in A$, it is called a monoid. If for every element $a$ of the monoid $A$, there exists an inverse element $a^{-1}\in A$ such that $aa^{-1}=a^{-1}a=e$, the monoid is called a group. A group $(A,\cdot)$ with commutative law i.e.
$$ab=ba,$$
for any $a,b\in A$ is called an abelian group named after a Norwegian mathematician Niels Abel. Note the inverse ${}^{-1}$ can be regarded as an operation on $A$, a unary operation ${}^{-1}: A\longrightarrow A$ defined by ${}^{-1}(a)=a^{-1}$ for each $a\in A$. The identity element $e$ can be also regarded as an operation, a nullary operation $e:\{\varnothing\}\longrightarrow A$. Thus, formally a group can be written as $(A,\cdot,{}^{-1},e)$, a quadrupple of a nonempty set, a binary operation, a unary operation, and a nullary operation.

Now we know what a group is and apparently, group theory is the study of groups. But what exactly are we studying there? What I am about to say is not really limited to group theory but commonly applies to studying other algebraic structures as well. There are briefly two main objectives with studying groups. One is the classification of groups. This becomes particularly interesting with groups of finite order. Here the order of a group means the number of elements of a group. We would like to answer the question “how many different groups of order $n$ are there for each $n$ and what are they?” The classification gets harder as $n$ gets larger. There are groups with the same order that appear to be different. But don’t be decieved by the appearance. They may actually be the same group. What do we mean by same here? We say two groups of the same order same if there is a one-to-one and onto map (a bijection) that preserves operations. Such a map is called an isomorphism. It turns out that if a map $\psi: G\longrightarrow G’$ from a group $G$ to another group $G’$ preserves binary operation, it automatically preserves unary and nullary operations. Here we mean preserving binary operation by
$$\psi(ab)=\psi(a)\psi(b)$$
for any $a,b\in G$. If you have taken linear algebra (and I believe you have), you would notice that a linear map is a map that preserves vector addition and scalar multiplication. A map $\psi: G\longrightarrow G’$ which preserves binary operation is called a homomorphism. If a homomorphism $\psi: G\longrightarrow G’$ is one-to-one and onto, it is an isomorphism. An isomorphism $\psi: G\longrightarrow G$ from a group $G$ onto itself is called an automorphism. In group theory, if there is an isomorphism from a group to another group, we do not distinguish them no matter how different they appear to look. The other objective is to discover new groups from old groups. Some of the new groups may be smaller in size than the old ones. Here we mean smaller in size by having a smaller number of elements i.e. having a lesser order. Some examples are subgroups and quotient groups. Some of the new groups are larger in size than the old ones. An example is direct products. Subgroups, quotient groups (also called factor groups), direct products are the things we will study as means to get new groups from old groups.

Group theory has a significance in geometry. In geometry, symmetry plays an important role. There are different types of symmetries: reflections, rotations, and translations. An interesting connection between geometry and group theory is that these symmetries form groups (symmetry groups). The most general symmetry group of finite order is called a symmetric group. In mathematics, the embedding theorem is conceptually and philosophically important (though it may be practically less important). When we study mathematics, we often feel that the structures we study are highly abstract and we feel like they only exist in our consciousness but not in the physical world. The embedding theorem tells that those abstract structures we study are indeed substructures of a larger structure that we are familiar with in the physical world. The embedding theorem implicates that we are not making up those abstract mathematical structures but we are merely discovering them which already exist in the universe. This kind of view point is called Mathematical Platonism. It turns out that there is an embedding theorem in finite group theory, namely every group of finite order is a subgroup of a symmetric group. The embedding theorem is called Cayley theorem. This means that the study of finite groups boils down to studying symmetric groups.

Remark. There is a mathematical structure called algebras over field $K$ (usually $K=\mathbb{R}$ or $K=\mathbb{C}$). An algebra $\mathcal{A}$ over field $K$ is a vector space over $K$ with a product $\cdot:\mathcal{A}\times\mathcal{A}\longrightarrow\mathcal{A}$ which is distributive over addition:
$$a(b+c)=ab+ac,\ (a+b)c=ac+bc,\ \forall\ a,b,c\in\mathcal{A}.$$
(Here, the symbol $\forall$ is a logical symbol which has meaning “for each”, “for any”, “for every”, or “for all” depending on the context. I will talk more about logical symbols next time as I will use them often.) Note that an algebra $\mathcal{A}$ over field $K$ is not an algebra because the scalar product is not an operation on $\mathcal{A}$. The scalar product is in fact an action of the multiplicative group $K\setminus\{0\}$ on $\mathcal{A}$. Algebras over field $K$ are important structures in functional analysis.