--- /dev/null
+---
+project: 'dmaap-messagerouter-msgrtr'
+project_creation_date: '2017-08-29'
+lifecycle_state: 'Incubation'
+project_lead: &onap_releng_ptl
+ name: 'Ram Koya'
+ email: 'rk541m@att.com'
+ id: 'rampi_k'
+ company: 'ATT'
+ timezone: 'America/Dallas'
+primary_contact: *onap_releng_ptl
+issue_tracking:
+ type: 'jira'
+ url: 'https://jira.onap.org/projects/DMAAP'
+ key: 'DMAAP'
+meetings:
+ - type: 'zoom'
+ agenda: ''
+ url: 'https://wiki.onap.org/pages/viewpage.action?pageId=13599275'
+ server: 'n/a'
+ channel: 'n/a'
+ repeats: 'weekly'
+ time: '13:00 UTC'
+committers:
+ - <<: *onap_releng_ptl
+ - name: 'Ramdas Sawant'
+ email: 'rs873m@att.com'
+ company: 'ATT'
+ id: 'rs873m'
+ timezone: 'America/Dallas'
+ - name: 'Varun Gudisena'
+ email: 'vg411h@att.com'
+ company: 'ATT'
+ id: 'vg411h'
+ timezone: 'America/Dallas'
+tsc:
+ approval: 'https://lists.onap.org/pipermail/onap-tsc'
+ changes:
+ - type: 'Removal'
+ name: 'Habib Madani'
+ name: 'Xinhui Li'
+ name: 'Jing Wang'
fId = id;
fCreateTimeMs = System.currentTimeMillis();
fLastTouch = fCreateTimeMs;
- fPendingMsgs = new LinkedBlockingQueue<ConsumerRecord<String, String>>();
+ fPendingMsgs = new LinkedBlockingQueue<>();
fLogTag = fGroup + "(" + fId + ")/" + fTopic;
offset = 0;
state = Kafka011Consumer.State.OPENED;
public synchronized Consumer.Message nextMessage() {
try {
- if (fPendingMsgs.size() > 0) {
+ if (fPendingMsgs.isEmpty()) {
return makeMessage(fPendingMsgs.take());
}
} catch (InterruptedException x) {
log.warn("After size>0, pending msg take() threw InterruptedException. Ignoring. (" + x.getMessage() + ")",
x);
+ Thread.currentThread().interrupt();
}
Callable<Boolean> run = new Callable<Boolean>() {
}
} catch (KafkaException x) {
- log.debug(fLogTag + ": KafkaException " + x.getMessage());
+ log.debug(fLogTag + ": KafkaException ", x);
} catch (java.lang.IllegalStateException | java.lang.IllegalArgumentException x) {
- log.error(fLogTag + ": Illegal state/arg exception in Kafka consumer; dropping stream. "
- + x.getMessage());
+ log.error(fLogTag + ": Illegal state/arg exception in Kafka consumer; dropping stream. ", x);
}
future.get(consumerPollTimeOut, TimeUnit.SECONDS); // wait 1
// second
} catch (TimeoutException ex) {
+ log.error("TimeoutException in in Kafka consumer ", ex);
// timed out. Try to stop the code if possible.
String apiNodeId = null;
try {
apiNodeId = InetAddress.getLocalHost().getCanonicalHostName() + ":" + CambriaConstants.kDefault_Port;
} catch (UnknownHostException e1) {
- // TODO Auto-generated catch block
- log.error("unable to get the localhost address");
+ log.error("unable to get the localhost address ", e1);
}
try {
if (fKafkaLiveLockAvoider != null)
fKafkaLiveLockAvoider.unlockConsumerGroup(apiNodeId, fTopic + "::" + fGroup);
} catch (Exception e) {
- log.error("unlockConsumerGroup(" + apiNodeId + "," + fTopic + "::" + fGroup);
+ log.error("Exception in unlockConsumerGroup(" + apiNodeId + "," + fTopic + "::" + fGroup, e);
}
forcePollOnConsumer();
future.cancel(true);
} catch (Exception ex) {
+ log.error("Exception in in Kafka consumer ", ex);
// timed out. Try to stop the code if possible.
future.cancel(true);
}
// second
} catch (TimeoutException ex) {
// timed out. Try to stop the code if possible.
- log.info("Timeout Occured - Kafka connection closure with in 300 seconds by a Executors task");
+ log.info("Timeout Occured - Kafka connection closure with in 300 seconds by a Executors task ", ex);
future.cancel(true);
setState(Kafka011Consumer.State.OPENED);
} catch (Exception ex) {
// timed out. Try to stop the code if possible.
- log.error("Exception occured Occured - Kafka connection closure with in 300 seconds by a Executors task"
- + ex);
+ log.error("Exception Occured - Kafka connection closure with in 300 seconds by a Executors task ", ex);
future.cancel(true);
setState(Kafka011Consumer.State.OPENED);
return false;
}
} catch (Exception e) {
- log.error("Failed and go to Exception block for " + fGroup + " " + e.getMessage());
+ log.error("Failed and go to Exception block for " + fGroup +" ", e);
}
}
});
}
} catch (java.util.ConcurrentModificationException e) {
- log.error("Error occurs for " + e);
+ log.error("Error occurs for ", e);
} catch (Exception e) {
- log.error("Failed and go to Exception block for " + group + " " + e.getMessage());
+ log.error("Failed and go to Exception block for " + group + " ", e);
}
}
});
curatorConsumerCache.close();
log.info("Curator client closed");
} catch (ZkInterruptedException e) {
- log.warn("Curator client close interrupted: " + e.getMessage());
+ log.warn("Curator client close interrupted: ", e);
} catch (IOException e) {
- log.warn("Error while closing curator PathChildrenCache for KafkaConsumerCache" + e.getMessage());
+ log.warn("Error while closing curator PathChildrenCache for KafkaConsumerCache ", e);
}
curatorConsumerCache = null;
log.info(" ^ deleted " + fBaseZkPath + "/" + key);
} catch (NoNodeException e) {
log.warn("A consumer was deleted from " + fApiId
- + "'s cache, but no Cambria API node had ownership of it in ZooKeeper");
+ + "'s cache, but no Cambria API node had ownership of it in ZooKeeper ", e);
} catch (Exception e) {
- log.debug("Unexpected exception while deleting consumer: " + e.getMessage());
- log.info(" %%%%%%@# Unexpected exception while deleting consumer: " + e.getMessage());
+ log.debug("Unexpected exception while deleting consumer: ", e);
+ log.info(" %%%%%%@# Unexpected exception while deleting consumer: ", e);
}
try {
try {
curator.setData().forPath(consumerPath, fApiId.getBytes());
} catch (KeeperException.NoNodeException e) {
+ log.info("KeeperException.NoNodeException occured", e);
curator.create().creatingParentsIfNeeded().forPath(consumerPath, fApiId.getBytes());
}
log.info(fApiId + " successfully claimed ownership of consumer " + consumerKey);
// final KeyedMessage<String, String> data = new
// KeyedMessage<String, String>(topic, m.getKey(),
- // kms.add(data);
+
final ProducerRecord<String, String> data = new ProducerRecord<String, String>(topic, m.getKey(),
m.getMessage());
+ batchId + "]");
try {
// ctx.getConfigReader().getfPublisher().sendBatchMessage(topic,
- // kms);
+
ctx.getConfigReader().getfPublisher().sendBatchMessageNew(topic, pms);
// transactionLogs(batch);
for (message msg : batch) {
metricsSet.publishTick(sizeNow);
publishBatchCount = sizeNow;
count += sizeNow;
- // batchId++;
+
String endTime = sdf.format(new Date());
LOG.info("Batch End Details:[serverIp=" + ctx.getRequest().getLocalAddr() + ",Batch End Id="
+ batchId + ",Batch Total=" + publishBatchCount + ",Batch Start Time=" + startTime
+ batchId + "]");
try {
// ctx.getConfigReader().getfPublisher().sendBatchMessage(topic,
- // kms);
+
ctx.getConfigReader().getfPublisher().sendBatchMessageNew(topic, pms);
- // transactionLogs(batch);
+
for (message msg : batch) {
LogDetails logDetails = msg.getLogDetails();
LOG.info("Publisher Log Details : " + logDetails.getPublisherLogDetails());
pms.clear();
metricsSet.publishTick(sizeNow);
count += sizeNow;
- // batchId++;
+
String endTime = sdf.format(new Date());
publishBatchCount = sizeNow;
LOG.info("Batch End Details:[serverIp=" + ctx.getRequest().getLocalAddr() + ",Batch End Id=" + batchId
return logDetails;
}
- /*
- * public String getMetricsTopic() { return metricsTopic; }
- *
- * public void setMetricsTopic(String metricsTopic) { this.metricsTopic =
- * metricsTopic; }
- */
+
+
+
+
+
+