You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@pinot.apache.org by Pinot Slack Email Digest <sn...@apache.org> on 2020/07/08 02:00:07 UTC

Apache Pinot Daily Email Digest (2020-07-07)

<h3><u>#general</u></h3><br><strong>@alan.hardy: </strong>@alan.hardy has joined the channel<br><h3><u>#random</u></h3><br><strong>@alan.hardy: </strong>@alan.hardy has joined the channel<br><h3><u>#troubleshooting</u></h3><br><strong>@alan.hardy: </strong>@alan.hardy has joined the channel<br><strong>@somanshu.jindal: </strong>I am getting failures at the time of segment commit on controller and server keeps on retrying indefinitely. I have attached controller and server logs screenshot. (v0.3.0)
Table config:
```{
  "tableName": "transcript",
  "tableType": "REALTIME",
  "segmentsConfig": {
    "timeColumnName": "timestamp",
    "timeType": "MILLISECONDS",
    "schemaName": "transcript",
    "replicasPerPartition": "1"
  },
  "tenants": {},
  "tableIndexConfig": {
    "loadMode": "MMAP",
    "streamConfigs": {
      "streamType": "kafka",
      "stream.kafka.consumer.type": "lowlevel",
      "stream.kafka.topic.name": "transcript-topic",
      "stream.kafka.decoder.class.name": "org.apache.pinot.plugin.stream.kafka.KafkaJSONMessageDecoder",
      "stream.kafka.consumer.factory.class.name": "org.apache.pinot.plugin.stream.kafka20.KafkaConsumerFactory",
      "stream.kafka.broker.list": "localhost:9876",
      "realtime.segment.flush.threshold.time": "5m",
      "realtime.segment.flush.threshold.size": "5",
      "stream.kafka.consumer.prop.auto.offset.reset": "largest"
    }
  },
  "metadata": {
    "customConfigs": {}
  }
}```
schema
```{
  "schemaName": "transcript",
  "dimensionFieldSpecs": [
    {
      "name": "studentID",
      "dataType": "INT"
    },
    {
      "name": "firstName",
      "dataType": "STRING"
    },
    {
      "name": "lastName",
      "dataType": "STRING"
    },
    {
      "name": "gender",
      "dataType": "STRING"
    },
    {
      "name": "subject",
      "dataType": "STRING"
    }
  ],
  "metricFieldSpecs": [
    {
      "name": "score",
      "dataType": "FLOAT"
    }
  ],
  "dateTimeFieldSpecs": [{
    "name": "timestamp",
    "dataType": "LONG",
    "format" : "1:MILLISECONDS:EPOCH",
    "granularity": "1:MILLISECONDS"
  }]
}```<br><strong>@damianoporta: </strong>Hello everybody I have three servers where I want to put Zookeeper cluster and the other Pinot components. I wonder how can I create a single entry address/port to let other components connect to zookeeper using the same address and port. I did a similar thing for brokers, I set up a load balancer to always call the same address and port without knowing the ip of each broker. But I do not know setting a load balancer for zookeeper seems unnecessary.<br><strong>@elon.azoulay: </strong>Is it possible to add columns to a pinot table in place or do we have to save the data and recreate/reload?<br><strong>@pradeepgv42: </strong>Hi, is there a simple way to mention not for `REGEXP_LIKE`? (rather than trying to achieve that in the regular expression)<br><h3><u>#pinot-dev</u></h3><br><strong>@dlavoie: </strong>Morning team! I'm struggling in building local SNAPSHOT version of Pinot that I can refer from other project. the fact that the version are not explicitely defined but made configurable with properties `${revision}` and `${sha1}` (this is breaking maven conventions btw) makes it very hard. The maven release plugin jobs does a great job of updating versions numbers between releases, any specific reason we historically use this pattern? Maven doesn't recommend placeholders in version since you can't guarantee what was released vs what is in your current VCS. Cheers!<br><strong>@vallamsetty: </strong>@vallamsetty has joined the channel<br><h3><u>#presto-pinot-streaming</u></h3><br><strong>@elon.azoulay: </strong>Updated, let me know if you have any feedback<br><h3><u>#s3-multiple-buckets</u></h3><br><strong>@alan.hardy: </strong>@alan.hardy has joined the channel<br><strong>@alan.hardy: </strong>@alan.hardy has left the channel<br><strong>@somanshu.jindal: </strong>@somanshu.jindal has joined the channel<br>