Back to Glossary Index

Dagster Data Engineering Glossary:


Data Denoising

Remove noise or artifacts from data to improve its accuracy and quality.

Data denoising - a definition:

In the context of data pipelines, denoising refers to the process of removing noise from the data, which could be caused by various factors such as measurement errors, faulty sensors, or even human error during data entry. The goal is to extract useful information from the data and remove any unwanted signals that could interfere with the analysis.

Data denoising in Python:

Here are some practical examples of denoising techniques in Python:

Please note that you need to have the necessary Python libraries installed in your Python environment to run this code.

Moving Average Filter: This is a simple denoising technique that involves taking the average of a sliding window of data points to smooth out the noise. It can be implemented using the rolling function in pandas or convolve function in numpy.

import numpy as np

def moving_average_filter(data, window_size):
    window = np.ones(window_size)/float(window_size)
    convolved_data = np.convolve(data, window, 'same')
    print("Original data: {}".format(data))
    print("Moving average filtered data with window size {}: {}".format(window_size, convolved_data))
    return convolved_data

# example usage
data = np.array([1, 2, 3, 4, 5, 6, 7, 8, 9, 10])
window_size = 4
filtered_data = moving_average_filter(data, window_size)

This example will put yield the following output:

Original data: [ 1  2  3  4  5  6  7  8  9 10]
Moving average filtered data with window size 4: [0.75 1.5  2.5  3.5  4.5  5.5  6.5  7.5  8.5  6.75]

Median Filter: This technique involves replacing each data point with the median of its neighboring points. It can be implemented using the signal.medfilt function in scipy.

from scipy import signal

def median_filter(data, window_size):
    filtered_data = signal.medfilt(data, kernel_size=window_size)
    print("Original data: {}".format(data))
    print("Median filtered data with window size {}: {}".format(window_size, filtered_data))
    return filtered_data

# example usage
data = [1, 2, 3, 4, 5, 6, 7, 8, 9, 10]
window_size = 5
filtered_data = median_filter(data, window_size)

Which will output:

Original data: [1, 2, 3, 4, 5, 6, 7, 8, 9, 10]
Median filtered data with window size 5: [1 2 3 4 5 6 7 8 8 8]

Try changing the window_size value to 4 for an interesting result.

Since we are applying a median filter to an input signal using the medfilt() function from the scipy.signal module. The medfilt() function applies a median filter to the input signal using a sliding window with a specified size, and returns the filtered signal.

When the window size is even, such as when window_size = 4, the function raises an error because the medfilt() function requires an odd number for the window size. This is because the median filter operates by taking the median value of the data points within the window, and when the window size is even, the median value is not well-defined.

Wavelet Transform: This is a more advanced denoising technique that involves decomposing the signal into different frequency bands using wavelets and then thresholding the coefficients to remove the noise. It can be implemented using the pywt library (which may require installation using pip install PyWavelets.

import numpy as np
import pywt

def wavelet_denoising(data, wavelet='db4', level=1):
    coeffs = pywt.wavedec(data, wavelet, level=level)
    threshold = np.std(coeffs[-level])
    coeffs = [pywt.threshold(c, threshold) for c in coeffs]
    denoised_data = pywt.waverec(coeffs, wavelet)
    print("Original data: {}".format(data))
    print("Denoised data using wavelet {}: {}".format(wavelet, denoised_data))
    return denoised_data

# example usage
data = [1, 2, 3, 4, 5, 6, 7, 8, 9, 10]
wavelet = 'db4'
level = 1
denoised_data = wavelet_denoising(data, wavelet, level)

Which will yield the following output:

Original data: [1, 2, 3, 4, 5, 6, 7, 8, 9, 10]
Denoised data using wavelet db4: [0.95707075 2.00990332 2.94189799 3.95561577 5.00005885 5.96165328
 6.97197144 7.97478225 8.97266171 9.96551716]

These are just a few examples of denoising techniques that can be used in data pipelines. The choice of technique depends on the nature of the noise and the requirements of the analysis.


Other data engineering terms related to
Data Transformation:
Dagster Glossary code icon

Align

Aligning data can mean one of three things: aligning datasets, meeting business rules, or arranging data elements in memory.
An image representing the data engineering concept of 'Align'
Dagster Glossary code icon

Clean or Cleanse

Remove invalid or inconsistent data values, such as empty fields or outliers.
An image representing the data engineering concept of 'Clean or Cleanse'
Dagster Glossary code icon

Cluster

Group data points based on similarities or patterns to facilitate analysis and modeling.
An image representing the data engineering concept of 'Cluster'
Dagster Glossary code icon

Curate

Select, organize, and annotate data to make it more useful for analysis and modeling.
An image representing the data engineering concept of 'Curate'
Dagster Glossary code icon

Denormalize

Optimize data for faster read access by reducing the number of joins needed to retrieve related data.
An image representing the data engineering concept of 'Denormalize'
Dagster Glossary code icon

Derive

Extracting, transforming, and generating new data from existing datasets.
An image representing the data engineering concept of 'Derive'
Dagster Glossary code icon

Discretize

Transform continuous data into discrete categories or bins to simplify analysis.
An image representing the data engineering concept of 'Discretize'
Dagster Glossary code icon

ETL

Extract, transform, and load data between different systems.
An image representing the data engineering concept of 'ETL'
Dagster Glossary code icon

Encode

Convert categorical variables into numerical representations for ML algorithms.
An image representing the data engineering concept of 'Encode'
Dagster Glossary code icon

Filter

Extract a subset of data based on specific criteria or conditions.
An image representing the data engineering concept of 'Filter'
Dagster Glossary code icon

Fragment

Break data down into smaller chunks for storage and management purposes.
An image representing the data engineering concept of 'Fragment'
Dagster Glossary code icon

Homogenize

Make data uniform, consistent, and comparable.
An image representing the data engineering concept of 'Homogenize'
Dagster Glossary code icon

Impute

Fill in missing data values with estimated or imputed values to facilitate analysis.
An image representing the data engineering concept of 'Impute'
Dagster Glossary code icon

Linearize

Transforming the relationship between variables to make datasets approximately linear.
An image representing the data engineering concept of 'Linearize'

Munge

See 'wrangle'.
An image representing the data engineering concept of 'Munge'
Dagster Glossary code icon

Normalize

Standardize data values to facilitate comparison and analysis. Organize data into a consistent format.
Dagster Glossary code icon

Reduce

Convert a large set of data into a smaller, more manageable form without significant loss of information.
An image representing the data engineering concept of 'Reduce'
Dagster Glossary code icon

Reshape

Change the structure of data to better fit specific analysis or modeling requirements.
An image representing the data engineering concept of 'Reshape'
Dagster Glossary code icon

Serialize

Convert data into a linear format for efficient storage and processing.
An image representing the data engineering concept of 'Serialize'
Dagster Glossary code icon

Shred

Break down large datasets into smaller, more manageable pieces for easier processing and analysis.
Dagster Glossary code icon

Skew

An imbalance in the distribution or representation of data.
Dagster Glossary code icon

Split

Divide a dataset into training, validation, and testing sets for machine learning model training.
Dagster Glossary code icon

Standardize

Transform data to a common unit or format to facilitate comparison and analysis.
Dagster Glossary code icon

Tokenize

Convert data into tokens or smaller units to simplify analysis or processing.
An image representing the data engineering concept of 'Tokenize'

Transform

Convert data from one format or structure to another.
An image representing the data engineering concept of 'Transform'
Dagster Glossary code icon

Wrangle

Convert unstructured data into a structured format.
An image representing the data engineering concept of 'Wrangle'