How does Kapacitor handle duplicate points?


InfluxDB handles duplicates nicely, overwriting duplicate data points with the latest for that point in time. But how does Kapacitor, processing a stream, handle duplicates?

If a Kapacitor Task is reading a data stream and the stream contains multiple copies of the same piece of data for a given point in time, does it read it multiple times, or does it do like InfluxDB and read only the last one?


I found the answer: If Kapacitor sees a duplicate data point in its stream, it counts it as a separate data point. It doesn’t not filter out duplicates, like InfluxDB does.

Now I need to know how to filter out duplicate data points that come in our data stream… Anybody got any ideas?