LogoLogo
Back to OsmosDeveloper DocsOsmos BlogWhat's New
  • Welcome to Osmos
  • 👋Getting Started with Osmos
    • Terminology
  • 🎉What's New
  • 🧩Osmos API Reference
  • ⌨️Osmos Chat
  • 👩‍💻Developer Docs
    • Manage API Keys
    • Embedding an Osmos Uploader
    • Embedding Uploader Jobs Table
    • Turning on Advanced Mode Uploader
    • Customizing Uploader Styling
    • Passing Parameterized Fields
    • Configuring Uploader's "Recall" functionality
    • Optional Uploader Settings
    • Uploader Submission Callback
    • Configuring AutoClean for your Uploader
    • Uploader Client-Side Validation
      • Data Validators
      • Checking for Duplicate values in a field
      • Creating Dropdown-Controlled Fields
      • Dynamic Dropdown Options
      • Dropdown Interaction with Validation Functions
    • Validation and Transformation Webhooks
      • OpenAPI Validation Webhook Testing
    • Parser Webhook for file based connectors
  • 🔠Datasets
    • Osmos Datasets
      • Uploading Data to your Table
      • Creating Primary and Foreign keys
      • Osmos Dataset Destination Connector
      • Osmos Dataset Source Connector
      • Dataset Edits
    • Datasets Query Builder
      • Query Builder Metadata
    • Performing Look Ups
      • Performing Joins
        • Types of Joins
  • ⏏️Uploader
    • Creating an Osmos Uploader
      • Testing your Osmos Uploader
    • Uploader Validation Summary
    • Advanced Mode
      • Overview
      • Process
    • Standard Mode
      • Overview
      • AutoClean
      • Process
    • AI AutoMapping
    • Uploaders Page
    • Uploader Details Page
  • 🔀Pipelines
    • Step 1. Select the Source
    • Step 2. Select a Destination
    • Step 3. Map & Transform Data
    • Step 4. Schedule the Pipeline
    • Step 5. Review & Confirm
    • Pipelines Page
    • Pipeline Details Page
  • ⏩Data Transformations
    • AutoMap
    • Column Mapping & Data Cleanup Panel
    • QuickFixes
    • AI Value Mapping
    • AI AutoClean
    • Lookups
      • Performing Lookups
    • SmartFill
    • Formulas
      • Date & Time Formulas
        • DateTime Format Specifiers
        • Timezone specifiers
      • Math Formulas and Operators
      • Logical Formulas & Operators
        • True & False Casting
      • Text Formulas
      • Other Formulas
    • Deduplication
  • ↘️Source Connectors
    • Amazon S3
    • Azure Blob Storage
    • BigQuery
    • Email
    • FTP
    • Google Cloud Storage (GCS)
    • Google Drive
    • Google Sheets
    • HTTP API (Call an Osmos API)
    • HTTP API (Osmos Calls Your API)
    • Osmos Dataset
    • Snowflake
    • Accessing Sources behind firewall
  • ↖️Destination Connectors
    • Amazon S3
    • BigQuery
    • FTP
    • Google Cloud Storage (GCS)
    • Google Drive
    • Google Sheets
    • HTTP API (Call an Osmos API)
    • HTTP API (Osmos Calls Your API)
      • Passing Dynamic Tokens in the API Header
    • MySQL
    • Osmos Dataset
    • PostgreSQL
    • Snowflake
    • Accessing Destinations behind firewall
  • 🗂️Projects
  • ⚙️Administration
    • Email Notifications
  • 🔒Security
  • 📞Support
  • Back to Osmos.io
Powered by GitBook
On this page
  • Overview
  • Scenario
  • Features
  • Inputs
  • Build the Dataset and Tables
  • Configuring the Lookup (non-foreign key-based scenario)
  • Pipeline Configuration
  • Uploader Configuration

Was this helpful?

  1. Data Transformations
  2. Lookups

Performing Lookups

PreviousLookupsNextSmartFill

Last updated 9 months ago

Was this helpful?

Overview

Performing a lookup typically involves searching for a specific piece of information within a larger set of data or databases. In Osmos, once you have created your Dataset(s) and have added at least one Table, you are ready to perform a look-up.

A Lookup is created in the data transformation phase and is managed through a QuickFix.

Scenario

Customer ABC is ingesting customer data from various providers. Both the customer ID and customer name are required fields for their destination system. Unfortunately, the customer name is not included in the majority of the customer source data but all of the source data includes the customer ID. How are they able to populate the destination system with the customer name?

Features

  • Datasets

  • Pipelines and/or Uploader

Inputs

Two Dataset Tables are required.

  1. Customer List, in this scenario, we will call it Customer Master, and it includes the two required fields, customer ID and customer Name. Other fields can be included but won't be used for this scenario

  2. The ongoing data that will be uploaded and/or run through the pipeline will include at least one field to reference the data in the look-up. For this scenario, the Source Data will be called ABC Co Provider Monthly Detail. This data includes fields such as customer ID.

Build the Dataset and Tables

Step 1: Create the Dataset. Within that Dataset create a table for your Customer Master.

Step 2: Upload the Customer list into the Customer Master Table

For this scenario, the Dataset name is called Provider Information and the Table is called Customer Master.

Foreign key lookups are the safest way to ensure referential integrity, but note that lookups can be made based on any values within tables

Configuring the Lookup (non-foreign key-based scenario)

Pipeline Configuration

In a Pipeline configuration, the lookup is configured in the column mapping. The lookup is configured by way of a QuickFix.

Step 1: Map the Source, Customer ID field, to the Customer Name Destination Column

The source field is also called the match field in the QuickFix lookup.

In the column mapping, map the Source field, match field, to the desired look-up Destination Column field. The match field is required in the source field to trigger the Datasets Lookup QuickFix

Step 2: Create the Lookup

Once the column is mapped, go to QuickFixes, and select Datasets Lookup.

Configure the Lookup:

  1. Select the Dataset called Provider Information.

  2. Select the Table called Customer Master.

  3. Select the Lookup Field called Customer Name

  4. Select the Matcher Field called Customer ID

  5. Hit Submit

The result in the Destination Column will be the customer name.

Uploader Configuration

In a Pipeline configuration, the lookup is configured in the column mapping. The lookup is configured by way of a QuickFix.

Step 1: Map the Source, Customer ID field, to the Customer Name Destination Column

The source field is also called the match field in the QuickFix lookup.

In the column mapping, map the Source field, match field, to the desired look-up Destination Column field. The match field is required in the source field to trigger the Datasets Lookup QuickFix

Step 2: Create the Lookup

Once the column is mapped, go to QuickFixes, and select Datasets Lookup.

Configure the Lookup:

  1. Select the Dataset called Provider Information.

  2. Select the Table called Customer Master.

  3. Select the Lookup Field called Customer Name

  4. Select the Matcher Field called Customer ID

  5. Hit Submit

The result in the Destination Column will be the customer name.

For detailed steps, go .

⏩
here