From: Bartek Grzybowski Date: Tue, 24 Mar 2020 11:11:29 +0000 (+0100) Subject: Fix unused varbiales pylint warnings X-Git-Tag: 1.6.0~27 X-Git-Url: https://gerrit.onap.org/r/gitweb?p=demo.git;a=commitdiff_plain;h=3069ed97314d524296e31fab39833d4af9cf1e19 Fix unused varbiales pylint warnings Change-Id: I3f76a82d276c75f20453a9306b4f9964d90b5179 Issue-ID: INT-1482 Signed-off-by: Bartek Grzybowski --- diff --git a/tutorials/vFWDT/workflow/workflow.py b/tutorials/vFWDT/workflow/workflow.py index 8703be73..0b2b6039 100755 --- a/tutorials/vFWDT/workflow/workflow.py +++ b/tutorials/vFWDT/workflow/workflow.py @@ -159,7 +159,7 @@ def _get_aai_rel_link_data(data, related_to, search_key=None, match_dict=None): m_value = None rel_dict = data.get(rel_lst) if rel_dict: # check if data has relationship lists - for key, rel_list in rel_dict.items(): + for key, rel_list in rel_dict.items(): # pylint: disable=W0612 for rel in rel_list: if rel.get("related-to") == related_to: dval = None @@ -398,7 +398,7 @@ def _osdf_request(rancher_ip, onap_ip, aai_data, exclude, use_oof_cache): #print(json.dumps(template, indent=4)) with _no_ssl_verification(): - response = api.osdf.placement(body=template, params={}, headers={}) + response = api.osdf.placement(body=template, params={}, headers={}) # pylint: disable=W0612 #if response.body.get('error_message') is not None: # raise Exception(response.body['error_message']['explanation']) @@ -505,7 +505,7 @@ def _extract_has_appc_identifiers(has_result, demand, onap_ip): v_server['vserver-name'] = v_server['vserver-name'].replace("01", "02") hostname_cache.append(v_server['vserver-name']) - api = _init_python_aai_api(onap_ip) + api = _init_python_aai_api(onap_ip) # pylint: disable=W0612 vnfc_type = demand.lower() # with _no_ssl_verification(): # response = api.aai.vnfc(v_server['vserver-name'], body=None, params={}, headers={}) @@ -1142,7 +1142,7 @@ def _generate_cdt_payloads_for_vnf(vnf_info, vnfc_type, actions): for action_artifact in artifacts[action]: artifact_list.append({'artifact-name': action_artifact['name'], 'artifact-type': action_artifact['type']}) if action != 'AllAction': - req = _generate_cdt_artifact_request(req_id, action_artifact, action, vnfc_type) + req = _generate_cdt_artifact_request(req_id, action_artifact, action, vnfc_type) # pylint: disable=W0612 #print(json.dumps(req, indent=4)) #print(json.dumps(action_info, indent=4)) diff --git a/vnfs/DAaaS/microservices/PythonApps/python-kafkaConsumer-inference-app/src/producer/CustomKafkaProducer.py b/vnfs/DAaaS/microservices/PythonApps/python-kafkaConsumer-inference-app/src/producer/CustomKafkaProducer.py index 43303761..6de868b7 100644 --- a/vnfs/DAaaS/microservices/PythonApps/python-kafkaConsumer-inference-app/src/producer/CustomKafkaProducer.py +++ b/vnfs/DAaaS/microservices/PythonApps/python-kafkaConsumer-inference-app/src/producer/CustomKafkaProducer.py @@ -23,8 +23,7 @@ class CustomKafkaProducer: ) self.producer.flush() - except Exception as e: - #print("Error during producing to kafka topic. Stacktrace is %s",e) + except Exception: logging.error("Error during producing to kafka topic.") traceback.print_exc() diff --git a/vnfs/DAaaS/sample-apps/sample-minio-save-model/build/save-model-to-minio.py b/vnfs/DAaaS/sample-apps/sample-minio-save-model/build/save-model-to-minio.py index c5198471..91207640 100755 --- a/vnfs/DAaaS/sample-apps/sample-minio-save-model/build/save-model-to-minio.py +++ b/vnfs/DAaaS/sample-apps/sample-minio-save-model/build/save-model-to-minio.py @@ -49,7 +49,7 @@ if __name__ == "__main__": found = False try: - client.stat_object(bucket_name, model); + client.stat_object(bucket_name, model) found = True except Exception as err: found = False @@ -62,5 +62,5 @@ if __name__ == "__main__": if not found: try: client.fput_object(bucket_name, model, filepath, metadata=metadata) - except expression as identifier: + except Exception as identifier: print(err)