/** * ============LICENSE_START==================================================== * org.onap.aaf * =========================================================================== * Copyright (c) 2018 AT&T Intellectual Property. All rights reserved. * =========================================================================== * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. * ============LICENSE_END==================================================== * */ package org.onap.aaf.auth.dao.cass; import java.io.ByteArrayOutputStream; import java.io.DataInputStream; import java.io.DataOutputStream; import java.io.IOException; import java.nio.ByteBuffer; import java.util.HashMap; import java.util.HashSet; import java.util.Iterator; import java.util.List; import java.util.Map; import java.util.Map.Entry; import org.onap.aaf.auth.dao.Bytification; import org.onap.aaf.auth.dao.Cached; import org.onap.aaf.auth.dao.CassAccess; import org.onap.aaf.auth.dao.CassDAOImpl; import org.onap.aaf.auth.dao.Loader; import org.onap.aaf.auth.dao.Streamer; import org.onap.aaf.auth.env.AuthzTrans; import org.onap.aaf.auth.layer.Result; import org.onap.aaf.misc.env.APIException; import org.onap.aaf.misc.env.Env; import org.onap.aaf.misc.env.TimeTaken; import java.util.Set; import com.datastax.driver.core.Cluster; import com.datastax.driver.core.ResultSet; import com.datastax.driver.core.Row; import com.datastax.driver.core.exceptions.DriverException; /** * NsDAO * * Data Access Object for Namespace Data * * @author Jonathan * */ public class NsDAO extends CassDAOImpl { public static final String TABLE = "ns"; public static final String TABLE_ATTRIB = "ns_attrib"; public static final int CACHE_SEG = 0x40; // yields segment 0x0-0x3F public static final int ROOT = 1; public static final int COMPANY=2; public static final int APP = 3; private static final String BEGIN_BATCH = "BEGIN BATCH\n"; private static final String APPLY_BATCH = "\nAPPLY BATCH;\n"; private static final String SQSCCR = "';\n"; private static final String SQCSQ = "','"; private HistoryDAO historyDAO; private CacheInfoDAO infoDAO; private PSInfo psNS; public NsDAO(AuthzTrans trans, Cluster cluster, String keyspace) throws APIException, IOException { super(trans, NsDAO.class.getSimpleName(),cluster,keyspace,Data.class,TABLE, readConsistency(trans,TABLE), writeConsistency(trans,TABLE)); init(trans); } public NsDAO(AuthzTrans trans, HistoryDAO hDAO, CacheInfoDAO iDAO) throws APIException, IOException { super(trans, NsDAO.class.getSimpleName(),hDAO,Data.class,TABLE, readConsistency(trans,TABLE), writeConsistency(trans,TABLE)); historyDAO=hDAO; infoDAO = iDAO; init(trans); } ////////////////////////////////////////// // Data Definition, matches Cassandra DM ////////////////////////////////////////// private static final int KEYLIMIT = 1; /** * Data class that matches the Cassandra Table "role" * * @author Jonathan */ public static class Data extends CacheableData implements Bytification { public String name; public int type; public String description; public String parent; public Map attrib; // //////////////////////////////////////// // // Getters public Map attrib(boolean mutable) { if (attrib == null) { attrib = new HashMap<>(); } else if (mutable && !(attrib instanceof HashMap)) { attrib = new HashMap<>(attrib); } return attrib; } @Override public int[] invalidate(Cached cache) { return new int[] { seg(cache,name) }; } public NsSplit split(String name) { return new NsSplit(this,name); } @Override public ByteBuffer bytify() throws IOException { ByteArrayOutputStream baos = new ByteArrayOutputStream(); NSLoader.deflt.marshal(this,new DataOutputStream(baos)); return ByteBuffer.wrap(baos.toByteArray()); } @Override public void reconstitute(ByteBuffer bb) throws IOException { NSLoader.deflt.unmarshal(this,toDIS(bb)); } @Override public String toString() { return name; } } private void init(AuthzTrans trans) throws APIException, IOException { // Set up sub-DAOs if (historyDAO==null) { historyDAO = new HistoryDAO(trans, this); } if (infoDAO==null) { infoDAO = new CacheInfoDAO(trans,this); } String[] helpers = setCRUD(trans, TABLE, Data.class, NSLoader.deflt,4/*need to skip attrib */); psNS = new PSInfo(trans, SELECT_SP + helpers[FIELD_COMMAS] + " FROM " + TABLE + " WHERE parent = ?", new NSLoader(1),readConsistency); } private static final class NSLoader extends Loader implements Streamer { public static final int MAGIC=250935515; public static final int VERSION=1; public static final int BUFF_SIZE=48; public static final NSLoader deflt = new NSLoader(KEYLIMIT); public NSLoader(int keylimit) { super(keylimit); } @Override public Data load(Data data, Row row) { // Int more efficient data.name = row.getString(0); data.type = row.getInt(1); data.description = row.getString(2); data.parent = row.getString(3); return data; } @Override protected void key(Data data, int idx, Object[] obj) { obj[idx]=data.name; } @Override protected void body(Data data, int _idx, Object[] obj) { int idx = _idx; obj[idx]=data.type; obj[++idx]=data.description; obj[++idx]=data.parent; } @Override public void marshal(Data data, DataOutputStream os) throws IOException { writeHeader(os,MAGIC,VERSION); writeString(os, data.name); os.writeInt(data.type); writeString(os,data.description); writeString(os,data.parent); if (data.attrib==null) { os.writeInt(-1); } else { os.writeInt(data.attrib.size()); for (Entry es : data.attrib(false).entrySet()) { writeString(os,es.getKey()); writeString(os,es.getValue()); } } } @Override public void unmarshal(Data data, DataInputStream is) throws IOException { /*int version = */readHeader(is,MAGIC,VERSION); // If Version Changes between Production runs, you'll need to do a switch Statement, and adequately read in fields byte[] buff = new byte[BUFF_SIZE]; data.name = readString(is, buff); data.type = is.readInt(); data.description = readString(is,buff); data.parent = readString(is,buff); int count = is.readInt(); if (count>0) { Map da = data.attrib(true); for (int i=0;i create(AuthzTrans trans, Data data) { String ns = data.name; // Ensure Parent is set if (data.parent==null) { return Result.err(Result.ERR_BadData, "Need parent for %s", ns); } // insert Attributes StringBuilder stmt = new StringBuilder(); stmt.append(BEGIN_BATCH); attribInsertStmts(stmt, data); stmt.append(APPLY_BATCH); try { getSession(trans).execute(stmt.toString()); //// TEST CODE for Exception // boolean force = true; // if (force) { // throw new com.datastax.driver.core.exceptions.NoHostAvailableException(new HashMap<>()); //// throw new com.datastax.driver.core.exceptions.AuthenticationException(new InetSocketAddress(9999),"Sample Message"); // } ////END TEST CODE } catch (DriverException | APIException | IOException e) { reportPerhapsReset(trans,e); trans.info().log(stmt); return Result.err(Result.ERR_Backend, "Backend Access"); } return super.create(trans, data); } @Override public Result update(AuthzTrans trans, Data data) { String ns = data.name; // Ensure Parent is set if (data.parent==null) { return Result.err(Result.ERR_BadData, "Need parent for %s", ns); } StringBuilder stmt = new StringBuilder(); stmt.append(BEGIN_BATCH); try { Map localAttr = data.attrib; Result> rremoteAttr = readAttribByNS(trans,ns); if (rremoteAttr.notOK()) { return Result.err(rremoteAttr); } // update Attributes String str; for (Entry es : localAttr.entrySet()) { str = rremoteAttr.value.get(es.getKey()); if (str==null || !str.equals(es.getValue())) { attribUpdateStmt(stmt, ns, es.getKey(),es.getValue()); } } // No point in deleting... insert overwrites... // for (Entry es : remoteAttr.entrySet()) { // str = localAttr.get(es.getKey()); // if (str==null || !str.equals(es.getValue())) { // attribDeleteStmt(stmt, ns, es.getKey()); // } // } if (stmt.length()>BEGIN_BATCH.length()) { stmt.append(APPLY_BATCH); getSession(trans).execute(stmt.toString()); } } catch (DriverException | APIException | IOException e) { reportPerhapsReset(trans,e); trans.info().log(stmt); return Result.err(Result.ERR_Backend, CassAccess.ERR_ACCESS_MSG); } return super.update(trans,data); } /* (non-Javadoc) * @see org.onap.aaf.auth.dao.CassDAOImpl#read(com.att.inno.env.TransStore, java.lang.Object) */ @Override public Result> read(AuthzTrans trans, Data data) { Result> rld = super.read(trans, data); if (rld.isOKhasData()) { for (Data d : rld.value) { // Note: Map is null at this point, save time/mem by assignment Result> rabn = readAttribByNS(trans,d.name); if (rabn.isOK()) { d.attrib = rabn.value; } else { return Result.err(rabn); } } } return rld; } /* (non-Javadoc) * @see org.onap.aaf.auth.dao.CassDAOImpl#read(com.att.inno.env.TransStore, java.lang.Object[]) */ @Override public Result> read(AuthzTrans trans, Object... key) { Result> rld = super.read(trans, key); if (rld.isOKhasData()) { for (Data d : rld.value) { // Note: Map is null at this point, save time/mem by assignment Result> rabn = readAttribByNS(trans,d.name); if (rabn.isOK()) { d.attrib = rabn.value; } else { return Result.err(rabn); } } } return rld; } @Override public Result delete(AuthzTrans trans, Data data, boolean reread) { TimeTaken tt = trans.start("Delete NS Attributes " + data.name, Env.REMOTE); try { StringBuilder stmt = new StringBuilder(); attribDeleteAllStmt(stmt, data); try { getSession(trans).execute(stmt.toString()); } catch (DriverException | APIException | IOException e) { reportPerhapsReset(trans,e); trans.info().log(stmt); return Result.err(Result.ERR_Backend, CassAccess.ERR_ACCESS_MSG); } } finally { tt.done(); } return super.delete(trans, data, reread); } public Result> readAttribByNS(AuthzTrans trans, String ns) { Map map = new HashMap<>(); TimeTaken tt = trans.start("readAttribByNS " + ns, Env.REMOTE); try { ResultSet rs = getSession(trans).execute("SELECT key,value FROM " + TABLE_ATTRIB + " WHERE ns='" + ns + "';"); for (Iterator iter = rs.iterator();iter.hasNext(); ) { Row r = iter.next(); map.put(r.getString(0), r.getString(1)); } } catch (DriverException | APIException | IOException e) { reportPerhapsReset(trans,e); return Result.err(Result.ERR_Backend, CassAccess.ERR_ACCESS_MSG); } finally { tt.done(); } return Result.ok(map); } public Result> readNsByAttrib(AuthzTrans trans, String key) { Set set = new HashSet<>(); TimeTaken tt = trans.start("readNsBykey " + key, Env.REMOTE); try { ResultSet rs = getSession(trans).execute("SELECT ns FROM " + TABLE_ATTRIB + " WHERE key='" + key + "';"); for (Iterator iter = rs.iterator();iter.hasNext(); ) { Row r = iter.next(); set.add(r.getString(0)); } } catch (DriverException | APIException | IOException e) { reportPerhapsReset(trans,e); return Result.err(Result.ERR_Backend, CassAccess.ERR_ACCESS_MSG); } finally { tt.done(); } return Result.ok(set); } public Result attribAdd(AuthzTrans trans, String ns, String key, String value) { try { getSession(trans).execute(attribInsertStmt(new StringBuilder(),ns,key,value).toString()); return Result.ok(); } catch (DriverException | APIException | IOException e) { reportPerhapsReset(trans,e); return Result.err(Result.ERR_Backend, CassAccess.ERR_ACCESS_MSG); } } private StringBuilder attribInsertStmt(StringBuilder sb, String ns, String key, String value) { sb.append("INSERT INTO "); sb.append(TABLE_ATTRIB); sb.append(" (ns,key,value) VALUES ('"); sb.append(ns); sb.append(SQCSQ); sb.append(key); sb.append(SQCSQ); sb.append(value); sb.append("');"); return sb; } private StringBuilder attribUpdateStmt(StringBuilder sb, String ns, String key, String value) { sb.append("UPDATE "); sb.append(TABLE_ATTRIB); sb.append(" set value='"); sb.append(value); sb.append("' where ns='"); sb.append(ns); sb.append("' AND key='"); sb.append(key); sb.append("';"); return sb; } public Result attribRemove(AuthzTrans trans, String ns, String key) { try { getSession(trans).execute(attribDeleteStmt(new StringBuilder(),ns,key).toString()); return Result.ok(); } catch (DriverException | APIException | IOException e) { reportPerhapsReset(trans,e); return Result.err(Result.ERR_Backend, CassAccess.ERR_ACCESS_MSG); } } private StringBuilder attribDeleteStmt(StringBuilder stmt, String ns, String key) { stmt.append("DELETE FROM "); stmt.append(TABLE_ATTRIB); stmt.append(" WHERE ns='"); stmt.append(ns); stmt.append("' AND key='"); stmt.append(key); stmt.append("';"); return stmt; } private void attribDeleteAllStmt(StringBuilder stmt, Data data) { stmt.append(" DELETE FROM "); stmt.append(TABLE_ATTRIB); stmt.append(" WHERE ns='"); stmt.append(data.name); stmt.append(SQSCCR); } private void attribInsertStmts(StringBuilder stmt, Data data) { // INSERT new Attrib for (Entry es : data.attrib(false).entrySet() ) { stmt.append(" "); attribInsertStmt(stmt,data.name,es.getKey(),es.getValue()); } } /** * Add description to Namespace * @param trans * @param ns * @param description * @return */ public Result addDescription(AuthzTrans trans, String ns, String description) { try { getSession(trans).execute(UPDATE_SP + TABLE + " SET description = '" + description.replace("'", "''") + "' WHERE name = '" + ns + "';"); } catch (DriverException | APIException | IOException e) { reportPerhapsReset(trans,e); return Result.err(Result.ERR_Backend, CassAccess.ERR_ACCESS_MSG); } Data data = new Data(); data.name=ns; wasModified(trans, CRUD.update, data, "Added description " + description + " to namespace " + ns, null ); return Result.ok(); } public Result> getChildren(AuthzTrans trans, String parent) { return psNS.read(trans, R_TEXT, new Object[]{parent}); } /** * Log Modification statements to History * * @param modified which CRUD action was done * @param data entity data that needs a log entry * @param overrideMessage if this is specified, we use it rather than crafting a history message based on data */ @Override protected void wasModified(AuthzTrans trans, CRUD modified, Data data, String ... override) { boolean memo = override.length>0 && override[0]!=null; boolean subject = override.length>1 && override[1]!=null; //TODO Must log history HistoryDAO.Data hd = HistoryDAO.newInitedData(); hd.user = trans.user(); hd.action = modified.name(); hd.target = TABLE; hd.subject = subject ? override[1] : data.name; hd.memo = memo ? override[0] : (data.name + " was " + modified.name() + 'd' ); if (modified==CRUD.delete) { try { hd.reconstruct = data.bytify(); } catch (IOException e) { trans.error().log(e,"Could not serialize NsDAO.Data"); } } if (historyDAO.create(trans, hd).status!=Status.OK) { trans.error().log("Cannot log to History"); } if (infoDAO.touch(trans, TABLE,data.invalidate(cache)).notOK()) { trans.error().log("Cannot touch CacheInfo"); } } }