text
stringlengths 301
426
| source
stringclasses 3
values | __index_level_0__
int64 0
404k
|
---|---|---|
Robotics, Robots, Engineering, Technology.
(📷: ModLab) Depending on that objective, the robots will configure themselves into an efficient shape and/or uses objects in their vicinity to help them accomplish the task. If the system can’t find a solution to the problem, the robots will disassemble themselves until provided with new
|
medium
| 8,963 |
Artificial Intelligence, Machine Learning, Data Science, Data, Reinforcement Learning.
Photo by Testalize.me on Unsplash ReLLIE: Deep Reinforcement Learning for Customized Low-Light Image Enhancement(arXiv) Author : Rongkai Zhang, Lanqing Guo, Siyu Huang, Bihan Wen Abstract : Low-light image enhancement (LLIE) is a pervasive yet challenging problem, since: 1) low-light measurements
|
medium
| 8,965 |
Artificial Intelligence, Machine Learning, Data Science, Data, Reinforcement Learning.
may vary due to different imaging conditions in practice; 2) images can be enlightened subjectively according to diverse preferences by each individual. To tackle these two challenges, this paper presents a novel deep reinforcement learning based method, dubbed ReLLIE, for customized low-light
|
medium
| 8,966 |
Artificial Intelligence, Machine Learning, Data Science, Data, Reinforcement Learning.
enhancement. ReLLIE models LLIE as a markov decision process, i.e., estimating the pixel-wise image-specific curves sequentially and recurrently. Given the reward computed from a set of carefully crafted non-reference loss functions, a lightweight network is proposed to estimate the curves for
|
medium
| 8,967 |
Artificial Intelligence, Machine Learning, Data Science, Data, Reinforcement Learning.
enlightening of a low-light image input. As ReLLIE learns a policy instead of one-one image translation, it can handle various low-light measurements and provide customized enhanced outputs by flexibly applying the policy different times. Furthermore, ReLLIE can enhance real-world images with
|
medium
| 8,968 |
Artificial Intelligence, Machine Learning, Data Science, Data, Reinforcement Learning.
hybrid corruptions, e.g., noise, by using a plug-and-play denoiser easily. Extensive experiments on various benchmarks demonstrate the advantages of ReLLIE, comparing to the state-of-the-art methods. 2.Settling the Sample Complexity of Model-Based Offline Reinforcement Learning(arXiv) Author : Gen
|
medium
| 8,969 |
Artificial Intelligence, Machine Learning, Data Science, Data, Reinforcement Learning.
Li, Laixi Shi, Yuxin Chen, Yuejie Chi, Yuting Wei Abstract : This paper is concerned with offline reinforcement learning (RL), which learns using pre-collected data without further exploration. Effective offline RL would be able to accommodate distribution shift and limited data coverage. However,
|
medium
| 8,970 |
Artificial Intelligence, Machine Learning, Data Science, Data, Reinforcement Learning.
prior algorithms or analyses either suffer from suboptimal sample complexities or incur high burn-in cost to reach sample optimality, thus posing an impediment to efficient offline RL in sample-starved applications. We demonstrate that the model-based (or “plug-in”) approach achieves
|
medium
| 8,971 |
Artificial Intelligence, Machine Learning, Data Science, Data, Reinforcement Learning.
minimax-optimal sample complexity without burn-in cost for tabular Markov decisionprocesses (MDPs). Concretely, consider a finite-horizon (resp. γ-discounted infinite-horizon) MDP with S states and horizon H(resp. effective horizon 11−γ), and suppose the distribution shift of data is reflected by
|
medium
| 8,972 |
Artificial Intelligence, Machine Learning, Data Science, Data, Reinforcement Learning.
some single-policy clipped concentrability coefficient C⋆clipped. We prove that model-based offline RL yields ε-accuracy with a sample complexity of{H4SC⋆clippedε2(finite-horizon MDPs)SC⋆clipped(1−γ)3ε2(infinite-horizon MDPs)up to log factor, which is minimax optimal for the entire ε-range. Our
|
medium
| 8,973 |
Artificial Intelligence, Machine Learning, Data Science, Data, Reinforcement Learning.
algorithms are “pessimistic” variants of value iteration with Bernstein-style penalties, and do not require sophisticated variance reduction. 3.Building Decision Forest via Deep Reinforcement Learning(arXiv) Author : Guixuan Wen, Kaigui Wu Abstract : Ensemble learning methods whose base classifier
|
medium
| 8,974 |
Artificial Intelligence, Machine Learning, Data Science, Data, Reinforcement Learning.
is a decision tree usually belong to the bagging or boosting. However, no previous work has ever built the ensemble classifier by maximizing long-term returns to the best of our knowledge. This paper proposes a decision forest building method called MA-H-SAC-DF for binary classification via deep
|
medium
| 8,975 |
Artificial Intelligence, Machine Learning, Data Science, Data, Reinforcement Learning.
reinforcement learning. First, the building process is modeled as a decentralized partial observable Markov decisionprocess, and a set of cooperative agents jointly constructs all base classifiers. Second, the global state and local observations are defined based on informations of the parent node
|
medium
| 8,976 |
Artificial Intelligence, Machine Learning, Data Science, Data, Reinforcement Learning.
and the current location. Last, the state-of-the-art deep reinforcementmethod Hybrid SAC is extended to a multi-agent system under the CTDE architecture to find an optimal decision forest building policy. The experiments indicate that MA-H-SAC-DF has the same performance as random forest, Adaboost,
|
medium
| 8,977 |
Artificial Intelligence, Machine Learning, Data Science, Data, Reinforcement Learning.
and GBDT on balanced datasets and outperforms them on imbalanced datasets. 4.REPTILE: A Proactive Real-Time Deep Reinforcement Learning Self-adaptive Framework(arXiv) Author : Flavio Corradini, Miichele Loreti, Marco Piangerelli, Giacomo Rocchetti Abstract : In this work a general framework is
|
medium
| 8,978 |
Artificial Intelligence, Machine Learning, Data Science, Data, Reinforcement Learning.
proposed to support the development of software systems that are able to adapt their behaviour according to the operating environment changes. The proposed approach, named REPTILE, works in a complete proactive manner and relies on Deep Reinforcement Learning-based agents to react to events,
|
medium
| 8,979 |
Artificial Intelligence, Machine Learning, Data Science, Data, Reinforcement Learning.
referred as novelties, that can affect the expected behaviour of the system. In our framework, two types of novelties are taken into account: those related to the context/environment and those related to the physical architecture itself. The framework, predicting those novelties before their
|
medium
| 8,980 |
Artificial Intelligence, Machine Learning, Data Science, Data, Reinforcement Learning.
occurrence, extracts time-changing models of the environment and uses a suitable Markov Decision Process to deal with the real-time setting. Moreover, the architecture of our RL agent evolves based on the possible actions that can be taken. 5.Computationally efficient joint coordination of multiple
|
medium
| 8,981 |
Artificial Intelligence, Machine Learning, Data Science, Data, Reinforcement Learning.
electric vehicle charging points using reinforcement learning(arXiv) Author : Manu Lahariya, Nasrin Sadeghianpourhamami, Chris Develder Abstract : A major challenge in todays power grid is to manage the increasing load from electric vehicle (EV) charging. Demand response (DR) solutions aim to
|
medium
| 8,982 |
Artificial Intelligence, Machine Learning, Data Science, Data, Reinforcement Learning.
exploit flexibility therein, i.e., the ability to shift EV charging in time and thus avoid excessive peaks or achieve better balancing. Whereas the majority of existing research works either focus on control strategies for a single EV charger, or use a multi-step approach (e.g., a first high level
|
medium
| 8,983 |
Artificial Intelligence, Machine Learning, Data Science, Data, Reinforcement Learning.
aggregate control decision step, followed by individual EV control decisions), we rather propose a single-step solution that jointly coordinates multiple charging points at once. In this paper, we further refine an initial proposal using reinforcement learning (RL), specifically addressing
|
medium
| 8,984 |
Artificial Intelligence, Machine Learning, Data Science, Data, Reinforcement Learning.
computational challenges that would limit its deployment in practice. More precisely, we design a new Markov decision process (MDP) formulation of the EV charging coordination process, exhibiting only linear space and time complexity (as opposed to the earlier quadratic space complexity). We thus
|
medium
| 8,985 |
Artificial Intelligence, Machine Learning, Data Science, Data, Reinforcement Learning.
improve upon earlier state-of-the-art, demonstrating 30% reduction of training time in our case study using real-world EV charging session data. Yet, we do not sacrifice the resulting performance in meeting the DR objectives: our new RL solutions still improve the performance of charging demand
|
medium
| 8,986 |
Machine Learning, Aritificial Intelligence, Data Science.
Euclidean space, the familiar geometry of our everyday world, provides a useful framework for understanding basic geometric concepts like angles and distances between points, lines, and planes. But in machine learning (ML), where data can be abstract, we need to venture beyond these familiar
|
medium
| 8,988 |
Machine Learning, Aritificial Intelligence, Data Science.
boundaries. To tackle the diverse challenges in ML, we turn to more abstract mathematical spaces that offer a richer framework for understanding and manipulating data. Feature spaces, where each data point is represented as a vector, are crucial for algorithms to identify patterns and make
|
medium
| 8,989 |
Machine Learning, Aritificial Intelligence, Data Science.
predictions. Metric spaces equip these feature spaces with a distance function, allowing for measuring similarity between data points. Hilbert spaces, a specific type of metric space with additional properties like completeness and an inner product, are particularly useful for kernel-based methods
|
medium
| 8,990 |
Machine Learning, Aritificial Intelligence, Data Science.
such as SVM. Their role is less prominent in deep learning, where complex non-linear models often rely more on empirical experimentation and intuition-driven design. However, this article seeks to demystify the mathematical terminology prevalent in research papers that utilize mathematical spaces,
|
medium
| 8,991 |
Machine Learning, Aritificial Intelligence, Data Science.
making the concepts more accessible and less intimidating for readers. Prerequisite This article assumes familiarity with the basic terms outlined in the previous article titled “The Study of Mathematical Spaces (Machine Learning)”. Metric Space While the concept of distance between two points is
|
medium
| 8,992 |
Machine Learning, Aritificial Intelligence, Data Science.
familiar, the power of metric spaces lies in extending this concept to less intuitive mathematical objects, allowing us to apply the properties and operations of metric spaces in broader contexts. To refresh our understanding, a metric space, denoted ⟨M, d⟩, is essentially a set M along with a
|
medium
| 8,993 |
Machine Learning, Aritificial Intelligence, Data Science.
well-defined distance function (metric) d. The metric d is a function that formalizes the notion of distance between elements of the set M. L1 and L2 distances are some popular examples for the metric in ML. To qualify as a metric function, three simple conditions must be satisfied. Non-negativity:
|
medium
| 8,994 |
Machine Learning, Aritificial Intelligence, Data Science.
𝑑(𝑥,𝑦)≥0. Symmetry: The distance is the same in both directions. Triangle Inequality: The direct route is the shortest. This level of abstraction offers considerable flexibility in defining a metric. Consequently, it grants substantial freedom to design metrics tailored to specific mathematical
|
medium
| 8,995 |
Machine Learning, Aritificial Intelligence, Data Science.
objects and problems. The goal is to find a metric that proves beneficial in solving your particular challenges. A notable example is the Wasserstein metric, which quantifies the similarity between two probability distributions and has gained popularity as a loss function in Generative Adversarial
|
medium
| 8,996 |
Machine Learning, Aritificial Intelligence, Data Science.
Networks (GANs). Complete Metric Spaces Completeness is a fundamental property of metric spaces. It means that every Cauchy sequence (a sequence where the terms get arbitrarily close together as they progress) converges to a limit that exists within the space. Formally, a metric space M is complete
|
medium
| 8,997 |
Machine Learning, Aritificial Intelligence, Data Science.
if every Cauchy sequence in M converges to a point that is also in M. This implies that M doesn’t have any “holes” or “missing points.” Completeness is crucial in various mathematical fields because it provides a foundation for analyzing limits and continuity. In machine learning, optimization
|
medium
| 8,998 |
Machine Learning, Aritificial Intelligence, Data Science.
algorithms like gradient descent rely on convergence properties. Theoretically, if the underlying space is complete, we can be confident that the algorithm will converge to a solution within the space, ensuring the results are meaningful and reliable. Examples Consider the sequence All elements in
|
medium
| 8,999 |
Machine Learning, Aritificial Intelligence, Data Science.
this sequence are real numbers inside the interval (0,1). But the limit, 1, is not in the interval (0, 1), and therefore the metric space ⟨X, d⟩ for X = (1, 0) with the usual metric 𝑑(𝑥, 𝑦) = |𝑥−𝑦| is not complete. To address this problem, we can expand X to [0, 1], making the metric space
|
medium
| 9,000 |
Machine Learning, Aritificial Intelligence, Data Science.
complete. To assess the completeness of a space, we first need to define a metric. The set of rational numbers equipped with the standard Euclidean metric is not complete. This is because a sequence of rational numbers can converge to an irrational number, which is not part of the rational number
|
medium
| 9,001 |
Machine Learning, Aritificial Intelligence, Data Science.
set. For example, even we can construct a Cauchy sequence of rational numbers ℚ that converges to the square root of 2. √2 is missing from ℚ. Vector Spaces (Linear Space) Vector spaces not only offer a natural structure to organize data points and features but also underpin linear models, such as
|
medium
| 9,002 |
Machine Learning, Aritificial Intelligence, Data Science.
linear regression and logistic regression, in many ML algorithms. Additionally, they facilitate the linear transformations central to model optimization. Their impact extends to kernel methods and natural language processing, where they enable the capture of semantic relationships through word
|
medium
| 9,003 |
Machine Learning, Aritificial Intelligence, Data Science.
embeddings. Vector spaces, also called linear spaces, provides a structure for manipulating objects called vectors. These vectors can be added together and multiplied (scaled) by numbers called scalars. Mathematically, a vector space over a field 𝐹 (commonly the real numbers ℝ or the complex
|
medium
| 9,004 |
Machine Learning, Aritificial Intelligence, Data Science.
numbers ℂ) is a set 𝑉 equipped with two operations: vector addition and scalar multiplication. The set 𝑉 and the operations must satisfy the following axioms, which hold for all elements 𝑢, 𝑣, w in V and all scalars 𝑎, 𝑏 in F: These operations are closed under V. The addition and the multiplication
|
medium
| 9,005 |
Machine Learning, Aritificial Intelligence, Data Science.
results remain under the same vector space. Vectors in a general vector space are abstract mathematical objects. They are not limited to the two-dimensional arrows used to represent physical quantities like force or velocity. The concept of a vector is much broader and can encompass many different
|
medium
| 9,006 |
Machine Learning, Aritificial Intelligence, Data Science.
types of objects. Example Polynomials, like 3x + 4x², x + 2x² + 0.5x⁵, and 5x², form a vector space V over the real numbers. We can add polynomials together and multiply them by real numbers (scalars), and the result will always be another polynomial within the same vector space V. For example, if
|
medium
| 9,007 |
Machine Learning, Aritificial Intelligence, Data Science.
we add (3x + 4x²) and (x + 2x² + 0.5x⁵), the result is 4x + 6x² + 0.5x⁵, which is also a polynomial in V. Therefore, functions can form a vector space, and each element in 𝑉 is simply a function. Normed Space & Banach Space A normed space is a vector space equipped with a norm, which measures the
|
medium
| 9,008 |
Machine Learning, Aritificial Intelligence, Data Science.
length or magnitude of its vectors. A Banach space is a complete normed space. Norm The norm of a vector v is written as ‖v‖. It is a function that assigns to each vector in a vector space V a non-negative real number (including zero), representing its length, magnitude, or size. The norm
|
medium
| 9,009 |
Machine Learning, Aritificial Intelligence, Data Science.
generalizes the familiar notion of length in Euclidean space to more abstract vector spaces. A normed vector space is a vector space defined over the field of real or complex numbers (𝔽 ∈ {ℝ, ℂ}). The norm maps each vector to a scalar in the field 𝔽. Any definition of ‖ · ‖ is called a norm if it
|
medium
| 9,010 |
Machine Learning, Aritificial Intelligence, Data Science.
satisfies the following properties: positive definiteness, absolute homogeneity, and the triangle inequality. These three rules also imply a norm is greater or equal to 0. This is the reason why we omit it from the norm axioms. Here are the common definitions for norms in ML. Normed Space (Normed
|
medium
| 9,011 |
Machine Learning, Aritificial Intelligence, Data Science.
Vector Space / Normed Linear Space) ⟨ X, ‖ · ‖ ⟩ forms a normed space. Here are two examples in the normed space: the first one uses an absolute value norm while the second one uses a p-norm. It’s important to note that a normed space is defined by its norm. The behavior of the space, including
|
medium
| 9,012 |
Machine Learning, Aritificial Intelligence, Data Science.
concepts like convergence and distance, depends on the specific norm being used. Different norms can lead to different conclusions about the properties of the space. Every normed space inherently possesses a metric induced by its norm. This allows norms to quantify both lengths of individual
|
medium
| 9,013 |
Machine Learning, Aritificial Intelligence, Data Science.
vectors and distances between vectors within the space. A normed space can simply use the norm to define its metric. Normed spaces are commonly used to quantify the similarity between data points, assess the quality of model predictions, and design optimization algorithms. For example, the L2 norm
|
medium
| 9,014 |
Machine Learning, Aritificial Intelligence, Data Science.
is frequently used to measure the error between predicted and true values in regression tasks, while the L1 norm promotes sparsity in feature selection. Banach Spaces A Banach space is a complete normed space (X, ‖ · ‖). Hilbert Space Hilbert spaces, as a specific type of complete normed vector
|
medium
| 9,015 |
Machine Learning, Aritificial Intelligence, Data Science.
space with additional structure, offer several advantages for ML applications: Inner Product: The defining feature of a Hilbert space is its inner product, which generalizes the dot product in Euclidean space. This inner product allows for the calculation of angles and projections, enabling
|
medium
| 9,016 |
Machine Learning, Aritificial Intelligence, Data Science.
techniques such as: Principal Component Analysis (PCA): Finding directions of maximal variance in data. Support Vector Machines (SVM): Finding optimal hyperplanes for classification. Orthogonality: Inner products facilitate the concept of orthogonality, which is crucial for: Fourier Transforms:
|
medium
| 9,017 |
Machine Learning, Aritificial Intelligence, Data Science.
Decomposing signals into orthogonal frequency components. Function Approximation: Representing functions as linear combinations of orthogonal basis functions. Completeness: Hilbert spaces are complete, meaning that every Cauchy sequence converges within the space. This is important for:
|
medium
| 9,018 |
Machine Learning, Aritificial Intelligence, Data Science.
Optimization: Guaranteeing the existence of solutions to optimization problems. Function Spaces: Many function spaces used in ML (e.g., spaces of square-integrable functions) are naturally Hilbert spaces. Duality: The Riesz representation theorem establishes a powerful duality between Hilbert
|
medium
| 9,019 |
Machine Learning, Aritificial Intelligence, Data Science.
spaces and their continuous dual spaces (spaces of linear functionals). This is exploited in: Kernel Methods: Optimizing the computation of similarity between two points projected in high-dimensional space using properties of the inner product space. Reproducing Kernel Hilbert Spaces (RKHS): A type
|
medium
| 9,020 |
Machine Learning, Aritificial Intelligence, Data Science.
of Hilbert space that provides a framework for kernel-based learning algorithms. Inner product The dot product is an example of an inner product, but the concept of an inner product is more abstract. An inner product is a binary operation on a vector space that satisfies the following axioms for
|
medium
| 9,021 |
Machine Learning, Aritificial Intelligence, Data Science.
all vectors x, y, z in the space and all scalars a: Inner Product Space & Hilbert Space A Hilbert space is endowed with an inner product and is complete. If the space is not complete, it is referred to as an inner product space. The inner product establishes a framework for abstract geometric
|
medium
| 9,022 |
Machine Learning, Aritificial Intelligence, Data Science.
concepts such as length, distance, and angle within a vector space. In every inner product space, we can derive a norm induced by the inner product: This norm allows us to define the distance between two vectors: Therefore, a Hilbert space is also a Banach space and a normed space. Furthermore, we
|
medium
| 9,023 |
Machine Learning, Aritificial Intelligence, Data Science.
can determine the angle between two vectors x and y: These definitions of norm, distance, and angle establish the geometric structure of inner product spaces. Both the set of real numbers (ℝ) and the set of 𝑛-dimensional real vectors (ℝⁿ) can form Hilbert spaces. In these cases, the inner product
|
medium
| 9,024 |
Machine Learning, Aritificial Intelligence, Data Science.
is typically defined as the dot product. Hilbert Space with Dot Product For Hilbert spaces consisting of functions, the inner product is often defined as an integral involving the product of the functions. Cauchy–Schwarz inequality The Cauchy-Schwarz inequality states that for all vectors x and y
|
medium
| 9,025 |
Machine Learning, Aritificial Intelligence, Data Science.
in an inner product space: Equality holds if and only if x and y are linearly dependent, meaning that one vector is a scalar multiple of the other. Orthogonality When the inner product of two vectors x and y is zero, they are said to be orthogonal, denoted as x ⊥ y. Orthogonal vectors are linearly
|
medium
| 9,026 |
Machine Learning, Aritificial Intelligence, Data Science.
independent, meaning neither can be expressed as a scalar multiple of the other. Similarly, functions f and g are orthogonal if their inner product is zero. A classic example is the pair of trigonometric functions sin(x) and cos(x), which are orthogonal over the interval [0, 2π]. In the broader
|
medium
| 9,027 |
Machine Learning, Aritificial Intelligence, Data Science.
context of vector spaces, any vector can be uniquely represented as a linear combination of vectors from an orthonormal basis. Notably, in function spaces, these basis vectors can themselves be functions. Subsets of a vector space X are considered independent if no element in one subset can be
|
medium
| 9,028 |
Machine Learning, Aritificial Intelligence, Data Science.
expressed as a linear combination of elements from the other subset. Use of Hilbert Space One notable application of Hilbert spaces in ML is the Reproducing Kernel Hilbert Space (RKHS). RKHS is a specialized type of Hilbert space that plays a crucial role in kernel-based learning algorithms,
|
medium
| 9,029 |
Machine Learning, Aritificial Intelligence, Data Science.
leveraging the inner product structure for efficient computation and representation of functions. The key advantage of RKHS is its ability to implicitly map data points into high-dimensional feature spaces, where similarities between them can be more easily found. Explicitly performing this mapping
|
medium
| 9,030 |
Machine Learning, Aritificial Intelligence, Data Science.
would be computationally expensive or even impossible in cases of infinite-dimensional feature spaces. However, the defining property of an RKHS, the reproducing property, combined with the kernel trick, enables efficient computation of these similarities using the kernel function without
|
medium
| 9,031 |
Machine Learning, Aritificial Intelligence, Data Science.
explicitly representing the data in the high-dimensional space. This is possible because the ultimate goal is often to compute scalar similarity values in the real numbers (ℝ), rather than the high-dimensional features themselves. 𝓁ᵖ Space The concept of infinity can be daunting in computer
|
medium
| 9,032 |
Machine Learning, Aritificial Intelligence, Data Science.
science. However, in ML, where algorithms often project data into high-dimensional spaces to uncover hidden patterns, it’s natural to consider extending this to infinite dimensions. While infinite-dimensional spaces might seem impractical, the key isn’t the number of dimensions, but whether we can
|
medium
| 9,033 |
Machine Learning, Aritificial Intelligence, Data Science.
efficiently compute relevant properties within those spaces. For example, if the sum of an infinite series or the limit of a sequence within an infinite-dimensional space is finite and computable, then infinite dimensionality can be a powerful tool for enhancing the expressiveness of ML models.
|
medium
| 9,034 |
Machine Learning, Aritificial Intelligence, Data Science.
This capability opens the door to studying and designing infinite-dimensional sequences. Elements in a mathematical space can be sequences. The ℓᵖ space is a set of infinite sequences where the p-th power of the absolute values of the terms is summable and finite. It is defined as the set of all
|
medium
| 9,035 |
Machine Learning, Aritificial Intelligence, Data Science.
infinite sequences x = (x₁, x₂, x₃, …) for which the infinite series formed by the p-th power of the absolute values of the elements converges. More formally, it is defined as: It is a normed space when endowed with the p-norm. To guarantee completeness, every Cauchy sequence in a space must
|
medium
| 9,036 |
Machine Learning, Aritificial Intelligence, Data Science.
converge to a limit within that space. In an ℓᵖ space, this condition is satisfied as the infinite series ∑|xᵢ|ᵖ converges for each sequence in the space. Therefore, ℓᵖ spaces qualify as Banach spaces, which are complete normed vector spaces. Banach Spaces (Complete Normed Space) Example Sequence
|
medium
| 9,037 |
Machine Learning, Aritificial Intelligence, Data Science.
in ℓ² Consider the following sequence as an example of an element in the ℓ² space: Inner Product for ℓ² Space The inner product in ℓ² can be defined as follows. Given the property below, the definition indeed represents an inner product in ℓ² space, fulfilling all necessary criteria. This will also
|
medium
| 9,038 |
Machine Learning, Aritificial Intelligence, Data Science.
serve as a good example to demonstrate how to prove that a function is an inner product. ℓ² pace & Hilbert Space The ℓ² space, often described as the space of square-summable sequences, is a prominent example of a Hilbert space, a type of complete inner product space. In ML, the ℓ² space, which is
|
medium
| 9,039 |
Machine Learning, Aritificial Intelligence, Data Science.
the space of square-summable sequences, is often used in the context of regularization techniques to prevent overfitting in models. ∞-norm When p = ∞, the corresponding norm is called the infinity norm (or supremum norm). Since there are infinitely many components in the sequence, finding a
|
medium
| 9,040 |
Machine Learning, Aritificial Intelligence, Data Science.
particular element as the maximum is not possible, so the infinity norm is computed using the supremum, which is the smallest value that is greater than or equal to all elements in the sequence. Lᵖ Space Lᵖ spaces are a specific type of function space defined for functions from a measurable space
|
medium
| 9,041 |
Machine Learning, Aritificial Intelligence, Data Science.
(often a subset of ℝⁿ) to ℝ (or ℂ). A measurable space extends the concept of size and volume to more abstract spaces. Infinite-dimensional spaces, like function spaces, allow us to model data as functions rather than just finite points. This concept is familiar to engineering students, who use
|
medium
| 9,042 |
Machine Learning, Aritificial Intelligence, Data Science.
Fourier transforms to represent data. In this context, the input is represented with periodic sinusoidal functions. The functions in Lᵖ spaces are required to be Lebesgue integrable, meaning the integral of the absolute value of the function raised to the power of p is finite. This ensures the
|
medium
| 9,043 |
Machine Learning, Aritificial Intelligence, Data Science.
completeness of the Lᵖ space. It is a normed vector space when endowed with the p-norm. The Gaussian function is in L² space. The Gaussian function 𝑓(𝑥) defined on the real line ℝ and specifically considered over an interval, say [−𝑎, 𝑎], is given by: Its square is integrable. The integral is
|
medium
| 9,044 |
Machine Learning, Aritificial Intelligence, Data Science.
finite for any finite a, because the exponential function rapidly decays to zero as 𝑥 moves away from zero, making the area under the curve finite. Indeed, the Gaussian function is in L² even over the entire real line (from -∞ to ∞). Inner Product for Lᵖ Space The inner product can be defined as:
|
medium
| 9,045 |
Machine Learning, Aritificial Intelligence, Data Science.
creating a Hilbert space. Continuous Function Space The continuous function space C[a, b] is defined as the set of all real-valued functions that are continuous on the closed interval [a, b]. In mathematical notation, this is expressed as: Let’s explore some visual examples of continuous functions
|
medium
| 9,046 |
Machine Learning, Aritificial Intelligence, Data Science.
within this space: Norm The corresponding p-norm is And the natural norm is defined as The metric d(u, v) below is a supremum between u(x) and v(x). (C, d) forms a complete metric space. Inner Product forms an inner product space. Examples of Hilbert Spaces Here is an recap of three possible
|
medium
| 9,047 |
Machine Learning, Aritificial Intelligence, Data Science.
Hilbert spaces, each defined by its unique set of elements and corresponding inner product: Euclidean space equipped with the dot product, ℓ² space characterized by the ℓ² norm, and the space of continuous functions on the interval [0,1]. Isomorphism Isomorphism is a structure-preserving mapping
|
medium
| 9,048 |
Machine Learning, Aritificial Intelligence, Data Science.
between two spaces. In the context of vector spaces, this preservation specifically refers to linearity, meaning that the operations of vector addition and scalar multiplication are maintained under the mapping. In a Banach space, an isomorphism preserves the norm following the mapping. That is,
|
medium
| 9,049 |
Machine Learning, Aritificial Intelligence, Data Science.
This is also referred to as an isometric isomorphism, where the distance is preserved before and after the mapping. For Hilbert spaces X and Y, and the mapping M: X → Y, isomorphism means that M is a bijective (one-to-one and onto) linear map that preserves the inner product, i.e., Recap Here is an
|
medium
| 9,050 |
Machine Learning, Aritificial Intelligence, Data Science.
overview of the relationships among some of the spaces we’ve studied. The arrows indicate the “is a” relationship. A normed vector space is inherently a metric space because the norm induces a metric, which quantifies the distance between vectors. Without a defined topology, a vector space by
|
medium
| 9,051 |
Web3, Blockchain Technology, Ontology Evm, Evm, Digital Asset Management.
EVM 101: Exploring Blockchain Technology and Decentralized Systems A primer on Virtual Machines and Ontology’s Ethereum Virtual Machine. In the rapidly evolving landscape of blockchain technology, interoperability and efficient digital asset management are not just goals — they’re necessities. As we
|
medium
| 9,053 |
Web3, Blockchain Technology, Ontology Evm, Evm, Digital Asset Management.
edge closer towards Web3, the next iteration of the internet, the dialogue around these topics has intensified, particularly the need to bridge different ecosystems and blockchains that currently operate in silos. This is where Virtual Machines (VMs), especially the Ethereum Virtual Machine (EVM),
|
medium
| 9,054 |
Web3, Blockchain Technology, Ontology Evm, Evm, Digital Asset Management.
come into play, playing a pivotal role in the seamless interaction between disparate networks. Ontology’s proactive approach, through the launch of the Ontology EVM, marks a significant milestone in enhancing cross-chain interoperability and bolstering the support for decentralized identity within
|
medium
| 9,055 |
Web3, Blockchain Technology, Ontology Evm, Evm, Digital Asset Management.
the Web3 space. This strategic move is not just about facilitating smoother transactions or interactions between blockchains; it’s about creating a more connected, accessible, and efficient digital world. The benefits of such integration extend far beyond simple connectivity, offering tangible
|
medium
| 9,056 |
Web3, Blockchain Technology, Ontology Evm, Evm, Digital Asset Management.
improvements in usability and accessibility for digital assets and decentralized identities. If the concept of VMs or the importance of interoperability in the Web3 ecosystem is new to you, don’t worry! We’ve got you covered. We’re recapping our series diving deep into the benefits of EVMs,
|
medium
| 9,057 |
Web3, Blockchain Technology, Ontology Evm, Evm, Digital Asset Management.
elucidating how they function as the backbone for interoperable and flexible blockchain networks. This series is a treasure trove of information for anyone looking to understand the critical role VMs play in the current and future landscape of blockchain technology. For a comprehensive overview, we
|
medium
| 9,058 |
Web3, Blockchain Technology, Ontology Evm, Evm, Digital Asset Management.
invite you to check out the post below for a full recap, ensuring you’re up to speed with how these technologies are shaping a more secure and interconnected digital future. VMs stand at the core of the blockchain revolution, not just as the backbone for deploying and managing decentralized
|
medium
| 9,059 |
Web3, Blockchain Technology, Ontology Evm, Evm, Digital Asset Management.
applications (dApps) but also as essential facilitators of smart contract interoperability across diverse blockchain ecosystems. These VMs offer an isolated environment where smart contracts — agreements expressed as code executed on the blockchain — can operate securely and without interference.
|
medium
| 9,060 |
Web3, Blockchain Technology, Ontology Evm, Evm, Digital Asset Management.
This secure execution environment ensures that data exchanges are safe and that the performance of applications is consistent, which is crucial for the reliability and trustworthiness of decentralized platforms. A smart contract’s programming language is typically specific to its blockchain
|
medium
| 9,061 |
Web3, Blockchain Technology, Ontology Evm, Evm, Digital Asset Management.
platform. However, the beauty of VMs lies in their ability to allow smart contracts written for one blockchain to be executed on another. This cross-chain functionality not only saves significant time and costs that would otherwise be spent on rewriting and testing new smart contracts for different
|
medium
| 9,062 |
Web3, Blockchain Technology, Ontology Evm, Evm, Digital Asset Management.
platforms but also eliminates the need for developers to learn new programming languages for each blockchain ecosystem they work with. By bridging these gaps, virtual machines serve as the linchpin of Web3, transforming decentralized platforms from theoretical concepts into practical solutions for
|
medium
| 9,063 |
Web3, Blockchain Technology, Ontology Evm, Evm, Digital Asset Management.
the digital challenges of today. They ensure that the blockchain ecosystem remains a vibrant and interconnected space, where the seamless exchange of digital assets and the smooth execution of decentralized identities are not just possible but are executed with enhanced usability and accessibility.
|
medium
| 9,064 |
Web3, Blockchain Technology, Ontology Evm, Evm, Digital Asset Management.
This positions VMs not merely as technical components but as critical enablers of a more secure, efficient, and interoperable digital future. For dApp developers, the journey from concept to launch is filled with challenges, innovation, and ultimately, satisfaction when seeing the number of users
|
medium
| 9,065 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.