Skip to content

Continue pipeline after invalid data #891

@hatamiarash7

Description

@hatamiarash7

I have a Kafka source with a huge amount of JSON data. This data was not validated before production and had many problems, such as non-UTF8 characters, bad integer data ( like "key": 004 ), corrupted JSON, etc.

Image

In the source configuration, I choose the Drop option to ignore bad data. But this will not work, and the processing is stopped after the first bad data.

As you can see in the screenshot, I have a stdout sink to see the output of the query. Everything works fine, but immediately after the first bad data, I get a warning, but the processing stops, and nothing is sent to the sink.

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions