Update project structure to org.onap.aaf
[aaf/authz.git] / authz-cass / src / main / java / org / onap / aaf / dao / CassDAOImpl.java
diff --git a/authz-cass/src/main/java/org/onap/aaf/dao/CassDAOImpl.java b/authz-cass/src/main/java/org/onap/aaf/dao/CassDAOImpl.java
new file mode 100644 (file)
index 0000000..61db914
--- /dev/null
@@ -0,0 +1,328 @@
+/*******************************************************************************\r
+ * ============LICENSE_START====================================================\r
+ * * org.onap.aaf\r
+ * * ===========================================================================\r
+ * * Copyright © 2017 AT&T Intellectual Property. All rights reserved.\r
+ * * ===========================================================================\r
+ * * Licensed under the Apache License, Version 2.0 (the "License");\r
+ * * you may not use this file except in compliance with the License.\r
+ * * You may obtain a copy of the License at\r
+ * * \r
+ *  *      http://www.apache.org/licenses/LICENSE-2.0\r
+ * * \r
+ *  * Unless required by applicable law or agreed to in writing, software\r
+ * * distributed under the License is distributed on an "AS IS" BASIS,\r
+ * * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\r
+ * * See the License for the specific language governing permissions and\r
+ * * limitations under the License.\r
+ * * ============LICENSE_END====================================================\r
+ * *\r
+ * * ECOMP is a trademark and service mark of AT&T Intellectual Property.\r
+ * *\r
+ ******************************************************************************/\r
+package org.onap.aaf.dao;\r
+\r
+import java.io.ByteArrayInputStream;\r
+import java.io.DataInputStream;\r
+import java.lang.reflect.Field;\r
+import java.nio.ByteBuffer;\r
+import java.util.List;\r
+\r
+import org.onap.aaf.authz.env.AuthzTrans;\r
+import org.onap.aaf.authz.layer.Result;\r
+import org.onap.aaf.dao.aaf.cass.Status;\r
+\r
+import org.onap.aaf.inno.env.TransStore;\r
+import com.datastax.driver.core.Cluster;\r
+import com.datastax.driver.core.ConsistencyLevel;\r
+import com.datastax.driver.core.ResultSet;\r
+import com.datastax.driver.core.ResultSetFuture;\r
+\r
+/**\r
+ * AbsCassDAO\r
+ *\r
+ * Deal with the essentials of Interaction with Cassandra DataStore for all Cassandra DAOs\r
+ *\r
+ *\r
+ * @param <DATA>\r
+ */\r
+public class CassDAOImpl<TRANS extends TransStore,DATA> extends AbsCassDAO<TRANS, DATA> implements DAO<TRANS,DATA> {\r
+       public static final String USER_NAME = "__USER_NAME__";\r
+       protected static final String CREATE_SP = "CREATE ";\r
+       protected static final String UPDATE_SP = "UPDATE ";\r
+       protected static final String DELETE_SP = "DELETE ";\r
+       protected static final String SELECT_SP = "SELECT ";\r
+\r
+       protected final String C_TEXT = getClass().getSimpleName() + " CREATE";\r
+       protected final String R_TEXT = getClass().getSimpleName() + " READ";\r
+       protected final String U_TEXT = getClass().getSimpleName() + " UPDATE";\r
+       protected final String D_TEXT = getClass().getSimpleName() + " DELETE";\r
+       private String table;\r
+       \r
+       protected final ConsistencyLevel readConsistency,writeConsistency;\r
+       \r
+       // Setteable only by CachedDAO\r
+       protected Cached<?, ?> cache;\r
+\r
+       /**\r
+        * A Constructor from the originating Cluster.  This DAO will open the Session at need,\r
+        * and shutdown the session when "close()" is called.\r
+        *\r
+        * @param cluster\r
+        * @param keyspace\r
+        * @param dataClass\r
+        */\r
+       public CassDAOImpl(TRANS trans, String name, Cluster cluster, String keyspace, Class<DATA> dataClass, String table, ConsistencyLevel read, ConsistencyLevel write) {\r
+               super(trans, name, cluster,keyspace,dataClass);\r
+               this.table = table;\r
+               readConsistency = read;\r
+               writeConsistency = write;\r
+       }\r
+       \r
+       /**\r
+        * A Constructor to share Session with other DAOs.\r
+        *\r
+        * This method get the Session and Cluster information from the calling DAO, and won't\r
+        * touch the Session on closure.\r
+        *\r
+        * @param aDao\r
+        * @param dataClass\r
+        */\r
+       public CassDAOImpl(TRANS trans, String name, AbsCassDAO<TRANS,?> aDao, Class<DATA> dataClass, String table, ConsistencyLevel read, ConsistencyLevel write) {\r
+               super(trans, name, aDao,dataClass);\r
+               this.table = table;\r
+               readConsistency = read;\r
+               writeConsistency = write;\r
+       }\r
+\r
+       protected PSInfo createPS;\r
+       protected PSInfo readPS;\r
+       protected PSInfo updatePS;\r
+       protected PSInfo deletePS;\r
+       private boolean async=false;\r
+\r
+       public void async(boolean bool) {\r
+               async = bool;\r
+       }\r
+\r
+       public final String[] setCRUD(TRANS trans, String table, Class<?> dc,Loader<DATA> loader) {\r
+               return setCRUD(trans, table, dc, loader, -1);\r
+       }\r
+       \r
+       public final String[] setCRUD(TRANS trans, String table, Class<?> dc,Loader<DATA> loader, int max) {\r
+                               Field[] fields = dc.getDeclaredFields();\r
+                               int end = max>=0 & max<fields.length?max:fields.length;\r
+                               // get keylimit from a non-null Loader\r
+                               int keylimit = loader.keylimit();\r
+                       \r
+                               StringBuilder sbfc = new StringBuilder();\r
+                               StringBuilder sbq = new StringBuilder();\r
+                               StringBuilder sbwc = new StringBuilder();\r
+                               StringBuilder sbup = new StringBuilder();\r
+                       \r
+                               if(keylimit>0) {\r
+                                       for(int i=0;i<end;++i) {\r
+                                               if(i>0) {\r
+                                                       sbfc.append(',');\r
+                                                       sbq.append(',');\r
+                                                       if(i<keylimit) {\r
+                                                               sbwc.append(" AND ");\r
+                                                       }\r
+                                               }\r
+                                               sbfc.append(fields[i].getName());\r
+                                               sbq.append('?');\r
+                                               if(i>=keylimit) {\r
+                                                       if(i>keylimit) {\r
+                                                               sbup.append(',');\r
+                                                       }\r
+                                                       sbup.append(fields[i].getName());\r
+                                                       sbup.append("=?");\r
+                                               }\r
+                                               if(i<keylimit) {\r
+                                                       sbwc.append(fields[i].getName());\r
+                                                       sbwc.append("=?");\r
+                                               }\r
+                                       }\r
+                       \r
+                                       createPS = new PSInfo(trans, "INSERT INTO " + table + " ("+ sbfc +") VALUES ("+ sbq +");",loader,writeConsistency);\r
+                       \r
+                                       readPS = new PSInfo(trans, "SELECT " + sbfc + " FROM " + table + " WHERE " + sbwc + ';',loader,readConsistency);\r
+                       \r
+                                       // Note: UPDATES can't compile if there are no fields besides keys... Use "Insert"\r
+                                       if(sbup.length()==0) {\r
+                                               updatePS = createPS; // the same as an insert\r
+                                       } else {\r
+                                               updatePS = new PSInfo(trans, "UPDATE " + table + " SET " + sbup + " WHERE " + sbwc + ';',loader,writeConsistency);\r
+                                       }\r
+                       \r
+                                       deletePS = new PSInfo(trans, "DELETE FROM " + table + " WHERE " + sbwc + ';',loader,writeConsistency);\r
+                               }\r
+                               return new String[] {sbfc.toString(), sbq.toString(), sbup.toString(), sbwc.toString()};\r
+                       }\r
+\r
+       public void replace(CRUD crud, PSInfo psInfo) {\r
+               switch(crud) {\r
+                       case create: createPS = psInfo; break;\r
+                       case read:   readPS = psInfo; break;\r
+                       case update: updatePS = psInfo; break;\r
+                       case delete: deletePS = psInfo; break;\r
+               }\r
+       }\r
+\r
+       public void disable(CRUD crud) {\r
+               switch(crud) {\r
+                       case create: createPS = null; break;\r
+                       case read:   readPS = null; break;\r
+                       case update: updatePS = null; break;\r
+                       case delete: deletePS = null; break;\r
+               }\r
+       }\r
+\r
+       \r
+       /**\r
+        * Given a DATA object, extract the individual elements from the Data into an Object Array for the\r
+        * execute element.\r
+        */\r
+       public Result<DATA> create(TRANS trans, DATA data)  {\r
+               if(createPS==null) {\r
+                       Result.err(Result.ERR_NotImplemented,"Create is disabled for %s",getClass().getSimpleName());\r
+               }\r
+               if(async) /*ResultSetFuture */ {\r
+                       Result<ResultSetFuture> rs = createPS.execAsync(trans, C_TEXT, data);\r
+                       if(rs.notOK()) {\r
+                               return Result.err(rs);\r
+                       }\r
+               } else {\r
+                       Result<ResultSet> rs = createPS.exec(trans, C_TEXT, data);\r
+                       if(rs.notOK()) {\r
+                               return Result.err(rs);\r
+                       }\r
+               }\r
+               wasModified(trans, CRUD.create, data);\r
+               return Result.ok(data);\r
+       }\r
+\r
+       /**\r
+        * Read the Unique Row associated with Full Keys\r
+        */\r
+       public Result<List<DATA>> read(TRANS trans, DATA data) {\r
+               if(readPS==null) {\r
+                       Result.err(Result.ERR_NotImplemented,"Read is disabled for %s",getClass().getSimpleName());\r
+               }\r
+               return readPS.read(trans, R_TEXT, data);\r
+       }\r
+\r
+       public Result<List<DATA>> read(TRANS trans, Object ... key) {\r
+               if(readPS==null) {\r
+                       Result.err(Result.ERR_NotImplemented,"Read is disabled for %s",getClass().getSimpleName());\r
+               }\r
+               return readPS.read(trans, R_TEXT, key);\r
+       }\r
+\r
+       public Result<Void> update(TRANS trans, DATA data) {\r
+               if(updatePS==null) {\r
+                       Result.err(Result.ERR_NotImplemented,"Update is disabled for %s",getClass().getSimpleName());\r
+               }\r
+               if(async)/* ResultSet rs =*/ {\r
+                       Result<ResultSetFuture> rs = updatePS.execAsync(trans, U_TEXT, data);\r
+                       if(rs.notOK()) {\r
+                               return Result.err(rs);\r
+                       }\r
+               } else {\r
+                       Result<ResultSet> rs = updatePS.exec(trans, U_TEXT, data);\r
+                       if(rs.notOK()) {\r
+                               return Result.err(rs);\r
+                       }\r
+               }\r
+               \r
+               wasModified(trans, CRUD.update, data);\r
+               return Result.ok();\r
+       }\r
+\r
+       // This method Sig for Cached...\r
+       public Result<Void> delete(TRANS trans, DATA data, boolean reread) {\r
+               if(deletePS==null) {\r
+                       Result.err(Result.ERR_NotImplemented,"Delete is disabled for %s",getClass().getSimpleName());\r
+               }\r
+               // Since Deleting will be stored off, for possible re-constitution, need the whole thing\r
+               if(reread) {\r
+                       Result<List<DATA>> rd = read(trans,data);\r
+                       if(rd.notOK()) {\r
+                               return Result.err(rd);\r
+                       }\r
+                       if(rd.isEmpty()) {\r
+                               return Result.err(Status.ERR_NotFound,"Not Found");\r
+                       }\r
+                       for(DATA d : rd.value) { \r
+                               if(async) {\r
+                                       Result<ResultSetFuture> rs = deletePS.execAsync(trans, D_TEXT, d);\r
+                                       if(rs.notOK()) {\r
+                                               return Result.err(rs);\r
+                                       }\r
+                               } else {\r
+                                       Result<ResultSet> rs = deletePS.exec(trans, D_TEXT, d);\r
+                                       if(rs.notOK()) {\r
+                                               return Result.err(rs);\r
+                                       }\r
+                               }\r
+                               wasModified(trans, CRUD.delete, d);\r
+                       }\r
+               } else {\r
+                       if(async)/* ResultSet rs =*/ {\r
+                               Result<ResultSetFuture> rs = deletePS.execAsync(trans, D_TEXT, data);\r
+                               if(rs.notOK()) {\r
+                                       return Result.err(rs);\r
+                               }\r
+                       } else {\r
+                               Result<ResultSet> rs = deletePS.exec(trans, D_TEXT, data);\r
+                               if(rs.notOK()) {\r
+                                       return Result.err(rs);\r
+                               }\r
+                       }\r
+                       wasModified(trans, CRUD.delete, data);\r
+               }\r
+               return Result.ok();\r
+       }\r
+       \r
+       public final Object[] keyFrom(DATA data) {\r
+               return createPS.keyFrom(data);\r
+       }\r
+\r
+       @Override\r
+       public String table() {\r
+               return table;\r
+       }\r
+       \r
+       public static final String CASS_READ_CONSISTENCY="cassandra.readConsistency";\r
+       public static final String CASS_WRITE_CONSISTENCY="cassandra.writeConsistency";\r
+       protected static ConsistencyLevel readConsistency(AuthzTrans trans, String table) {\r
+               String prop = trans.getProperty(CASS_READ_CONSISTENCY+'.'+table);\r
+               if(prop==null) {\r
+                       prop = trans.getProperty(CASS_READ_CONSISTENCY);\r
+                       if(prop==null) {\r
+                               return ConsistencyLevel.ONE; // this is Cassandra Default\r
+                       }\r
+               }\r
+               return ConsistencyLevel.valueOf(prop);\r
+       }\r
+\r
+       protected static ConsistencyLevel writeConsistency(AuthzTrans trans, String table) {\r
+               String prop = trans.getProperty(CASS_WRITE_CONSISTENCY+'.'+table);\r
+               if(prop==null) {\r
+                       prop = trans.getProperty(CASS_WRITE_CONSISTENCY);\r
+                       if(prop==null) {\r
+                               return ConsistencyLevel.ONE; // this is Cassandra Default\\r
+                       }\r
+               }\r
+               return ConsistencyLevel.valueOf(prop);\r
+       }\r
+\r
+       public static DataInputStream toDIS(ByteBuffer bb) {\r
+               byte[] b = bb.array();\r
+               return new DataInputStream(\r
+                       new ByteArrayInputStream(b,bb.position(),bb.limit())\r
+               );\r
+       }\r
+\r
+\r
+}\r