diff --git a/src/main/java/com/hy/java/uct/cdtocode/CDToCodeTracer.java b/src/main/java/com/hy/java/uct/cdtocode/CDToCodeTracer.java index cd6a0dcefad2dffe1b89a9fcd7097e9b383cf071..96c071f2955ba5328f90f4a679c8971ab85d46ab 100644 --- a/src/main/java/com/hy/java/uct/cdtocode/CDToCodeTracer.java +++ b/src/main/java/com/hy/java/uct/cdtocode/CDToCodeTracer.java @@ -48,11 +48,11 @@ public class CDToCodeTracer { /* * Hadoop HDFS */ - // Map classes_in_CD = CDReader.read(cd_dir + "cd-Hadoop HDFS.txt"); + Map classes_in_CD = CDReader.read(cd_dir + "cd-Hadoop HDFS.txt"); /* * Hadoop MapReduce */ - Map classes_in_CD = CDReader.read(cd_dir + "cd-Hadoop MapReduce.txt"); + // Map classes_in_CD = CDReader.read(cd_dir + "cd-Hadoop MapReduce.txt"); // 检查结果,可注释掉 // CDReader.check(classes_in_CD); /* @@ -77,35 +77,32 @@ public class CDToCodeTracer { /* * Hadoop HDFS */ + doc_dir_ls.add(doc_dir + "Hadoop HDFS\\Hadoop architectural overview.txt"); + doc_dir_ls.add(doc_dir + "Hadoop HDFS\\Hadoop clusters with Kove® XPD™ persistent memory.txt"); + doc_dir_ls.add(doc_dir + "Hadoop HDFS\\HADOOP DISTRIBUTED FILE SYSTEM (HDFS) ARCHITECTURAL DOCUMENTATION - MODULE VIEW.txt"); + doc_dir_ls.add(doc_dir + "Hadoop HDFS\\Hadoop Distributed File System (HDFS) Architecture – A Guide to HDFS for Every Data Engineer.txt"); + doc_dir_ls.add(doc_dir + "Hadoop HDFS\\HADOOP ECOSYSTEM.txt"); + doc_dir_ls.add(doc_dir + "Hadoop HDFS\\Hadoop HDFS Architecture Explanation and Assumptions.txt"); + doc_dir_ls.add(doc_dir + "Hadoop HDFS\\HDFS Architecture Guide.txt"); + doc_dir_ls.add(doc_dir + "Hadoop HDFS\\HDFS Architecture.txt"); + doc_dir_ls.add(doc_dir + "Hadoop HDFS\\HDFS.txt"); + doc_dir_ls.add(doc_dir + "Hadoop HDFS\\Key Design of HDFS Architecture.txt"); + doc_dir_ls.add(doc_dir + "Hadoop HDFS\\The Hadoop Distributed File System Architecture and Design.txt"); + doc_dir_ls.add(doc_dir + "Hadoop HDFS\\Towards A Scalable HDFS Architecture.txt"); /* - * doc_dir_ls.add(doc_dir + "Hadoop HDFS\\Hadoop architectural overview.txt"); doc_dir_ls.add(doc_dir + "Hadoop HDFS\\Hadoop clusters with Kove® XPD™ persistent memory.txt"); doc_dir_ls.add(doc_dir + - * "Hadoop HDFS\\HADOOP DISTRIBUTED FILE SYSTEM (HDFS) ARCHITECTURAL DOCUMENTATION - MODULE VIEW.txt"); doc_dir_ls.add(doc_dir + - * "Hadoop HDFS\\Hadoop Distributed File System (HDFS) Architecture – A Guide to HDFS for Every Data Engineer.txt"); doc_dir_ls.add(doc_dir + "Hadoop HDFS\\HADOOP ECOSYSTEM.txt"); - * doc_dir_ls.add(doc_dir + "Hadoop HDFS\\Hadoop HDFS Architecture Explanation and Assumptions.txt"); doc_dir_ls.add(doc_dir + "Hadoop HDFS\\HDFS Architecture Guide.txt"); doc_dir_ls.add(doc_dir + - * "Hadoop HDFS\\HDFS Architecture.txt"); doc_dir_ls.add(doc_dir + "Hadoop HDFS\\HDFS.txt"); doc_dir_ls.add(doc_dir + "Hadoop HDFS\\Key Design of HDFS Architecture.txt"); doc_dir_ls.add(doc_dir + - * "Hadoop HDFS\\The Hadoop Distributed File System Architecture and Design.txt"); doc_dir_ls.add(doc_dir + "Hadoop HDFS\\Towards A Scalable HDFS Architecture.txt"); + * Hadoop MapReduce */ /* - * Hadoop MapReduce + * doc_dir_ls.add(doc_dir + "Hadoop MapReduce\\A BIG DATA PROCESSING FRAMEWORK BASED ON MAPREDUCE WITH APPLICATION TO INTERNET OF THINGS.txt"); doc_dir_ls.add(doc_dir + + * "Hadoop MapReduce\\Apache Hadoop Architecture – HDFS, YARN & MapReduce.txt"); doc_dir_ls.add(doc_dir + "Hadoop MapReduce\\Big Data Analysis Challenges and Solutions.txt"); doc_dir_ls.add(doc_dir + + * "Hadoop MapReduce\\Big Data Management on Wireless Sensor Networks.txt"); doc_dir_ls.add(doc_dir + "Hadoop MapReduce\\Hadoop - MapReduce.txt"); doc_dir_ls.add(doc_dir + + * "Hadoop MapReduce\\Hadoop Architecture in Detail – HDFS, Yarn & MapReduce.txt"); doc_dir_ls.add(doc_dir + "Hadoop MapReduce\\Hadoop MapReduce- Java-based Processing Framework for Big Data.txt"); + * doc_dir_ls.add(doc_dir + "Hadoop MapReduce\\MapReduce – Components.txt"); doc_dir_ls.add(doc_dir + "Hadoop MapReduce\\MapReduce Architecture1.txt"); doc_dir_ls.add(doc_dir + + * "Hadoop MapReduce\\MapReduce Architecture2.txt"); doc_dir_ls.add(doc_dir + "Hadoop MapReduce\\MapReduce Architecture3.txt"); doc_dir_ls.add(doc_dir + "Hadoop MapReduce\\MapReduce Tutorial.txt"); + * doc_dir_ls.add(doc_dir + "Hadoop MapReduce\\MapReduce Working and Components.txt"); doc_dir_ls.add(doc_dir + "Hadoop MapReduce\\MapReduce.txt"); doc_dir_ls.add(doc_dir + + * "Hadoop MapReduce\\mapreduce_hadoop2.txt"); doc_dir_ls.add(doc_dir + "Hadoop MapReduce\\Understanding MapReduce in Hadoop.txt"); doc_dir_ls.add(doc_dir + + * "Hadoop MapReduce\\What are the components of MapReduce.txt"); doc_dir_ls.add(doc_dir + "Hadoop MapReduce\\What Is MapReduce Architecture An Important Overview For 2021.txt"); */ - doc_dir_ls.add(doc_dir + "Hadoop MapReduce\\A BIG DATA PROCESSING FRAMEWORK BASED ON MAPREDUCE WITH APPLICATION TO INTERNET OF THINGS.txt"); - doc_dir_ls.add(doc_dir + "Hadoop MapReduce\\Apache Hadoop Architecture – HDFS, YARN & MapReduce.txt"); - doc_dir_ls.add(doc_dir + "Hadoop MapReduce\\Big Data Analysis Challenges and Solutions.txt"); - doc_dir_ls.add(doc_dir + "Hadoop MapReduce\\Big Data Management on Wireless Sensor Networks.txt"); - doc_dir_ls.add(doc_dir + "Hadoop MapReduce\\Hadoop - MapReduce.txt"); - doc_dir_ls.add(doc_dir + "Hadoop MapReduce\\Hadoop Architecture in Detail – HDFS, Yarn & MapReduce.txt"); - doc_dir_ls.add(doc_dir + "Hadoop MapReduce\\Hadoop MapReduce- Java-based Processing Framework for Big Data.txt"); - doc_dir_ls.add(doc_dir + "Hadoop MapReduce\\MapReduce – Components.txt"); - doc_dir_ls.add(doc_dir + "Hadoop MapReduce\\MapReduce Architecture1.txt"); - doc_dir_ls.add(doc_dir + "Hadoop MapReduce\\MapReduce Architecture2.txt"); - doc_dir_ls.add(doc_dir + "Hadoop MapReduce\\MapReduce Architecture3.txt"); - doc_dir_ls.add(doc_dir + "Hadoop MapReduce\\MapReduce Tutorial.txt"); - doc_dir_ls.add(doc_dir + "Hadoop MapReduce\\MapReduce Working and Components.txt"); - doc_dir_ls.add(doc_dir + "Hadoop MapReduce\\MapReduce.txt"); - doc_dir_ls.add(doc_dir + "Hadoop MapReduce\\mapreduce_hadoop2.txt"); - doc_dir_ls.add(doc_dir + "Hadoop MapReduce\\Understanding MapReduce in Hadoop.txt"); - doc_dir_ls.add(doc_dir + "Hadoop MapReduce\\What are the components of MapReduce.txt"); - doc_dir_ls.add(doc_dir + "Hadoop MapReduce\\What Is MapReduce Architecture An Important Overview For 2021.txt"); // 实际使用的Map,保存每份文档地址及其内容 Map> dir_sentences_map = DocReader.readDocs(doc_dir_ls); /* @@ -115,8 +112,8 @@ public class CDToCodeTracer { */ // 记得改这里的路径 // Map classFullName_javaFileDir_map = CodeReader.read(code_dir + "code path-fm"); - // Map classFullName_javaFileDir_map = CodeReader.read(code_dir + "code path-hdfs"); - Map classFullName_javaFileDir_map = CodeReader.read(code_dir + "code path-mr"); + Map classFullName_javaFileDir_map = CodeReader.read(code_dir + "code path-hdfs"); + // Map classFullName_javaFileDir_map = CodeReader.read(code_dir + "code path-mr"); // 检查结果,可注释掉 // CodeReader.check(classFullName_javaFileDir_map); // 针对每个图中的类名,如果代码中有相似的包名,则将包下所有相似的类(字符串包含至少一个关键词即可)都重点标记。此脚本为一次性,使用后删除即可 @@ -142,19 +139,19 @@ public class CDToCodeTracer { * * 基于启发式模糊匹配的UML类图与代码追踪方法:首先针对类图中的类和代码中的类,基于类的名称进行字符串完全匹配,从而建立确定的初始追踪;基于同义词、 词缀词典等语料库,定义基于类名称匹配和关联关系的启发式匹配规则,研究基于模糊匹配技术的追踪关系建立方法,基于初始追踪和启发式规则, 对类名不一致的模型和代码元素进行启发式追踪,扩展初始追踪关系。 */ - Map mapped_classes = CodeRelationMapper.map(classShortName_classObj_mappedByDoc, classFullName_javaFileDir_map); + Map mapped_classes = CodeRelationMapper.map(classShortName_classObj_mappedByDoc, classFullName_javaFileDir_map, false); /* * Apache OODT File Manager */ - // CodeRelationMapper.save(mapped_classes, res_dir + "Apache OODT File Manager.xls"); + // CodeRelationMapper.save(mapped_classes, res_dir + "Apache OODT File Manager-noadd.xls"); /* * Hadoop HDFS */ - // CodeRelationMapper.save(mapped_classes, res_dir + "Hadoop HDFS.xls"); + CodeRelationMapper.save(mapped_classes, res_dir + "Hadoop HDFS-noadd.xls"); /* * Hadoop MapReduce */ - CodeRelationMapper.save(mapped_classes, res_dir + "Hadoop MapReduce.xls"); + // CodeRelationMapper.save(mapped_classes, res_dir + "Hadoop MapReduce-noadd.xls"); // 检查结果,可注释掉 // CodeRelationMapper.check(res_dir + "Apache OODT File Manager.xls"); } diff --git a/src/main/java/com/hy/java/uct/cdtocode/RelationTypeXlsReader.java b/src/main/java/com/hy/java/uct/cdtocode/RelationTypeXlsReader.java new file mode 100644 index 0000000000000000000000000000000000000000..fccb53ca52c4abceb3eb7b06a9ad3ada5a0fe615 --- /dev/null +++ b/src/main/java/com/hy/java/uct/cdtocode/RelationTypeXlsReader.java @@ -0,0 +1,89 @@ +package com.hy.java.uct.cdtocode; + +import java.io.File; +import java.io.IOException; + +import com.hy.java.utility.common.Traverser; +import com.hy.java.utility.common.Traverser.FileNode; + +import jxl.Sheet; +import jxl.Workbook; +import jxl.read.biff.BiffException; + +public class RelationTypeXlsReader { + /** + * 将设计文档放在doc_dir目录下 + */ + private static final String doc_dir = System.getProperty("user.dir") + "\\src\\main\\resources\\cdtocode\\doc\\"; + + public static void main(String[] args) { + FileNode srcdoc_fn = Traverser.traverseDir(doc_dir); + CountGroup cg = new CountGroup(); + for (FileNode proj_dir_fn : srcdoc_fn.children) { + if (!proj_dir_fn.path.contains("jetty")) { + for (FileNode doc_fn : proj_dir_fn.children) { + if (doc_fn.path.endsWith(".xls")) { + read(doc_fn.path, cg); + } + } + } + } + // System.out.println(cg.num_generalization); + // System.out.println(cg.num_association); + // System.out.println(cg.num_aggregation); + // System.out.println(cg.num_attribute); + System.out.println("========================================"); + double ratio_generalization = 100.0 * (double) cg.num_generalization / (double) (cg.num_generalization + cg.num_association + cg.num_aggregation + cg.num_attribute); + double ratio_association = 100.0 * (double) cg.num_association / (double) (cg.num_generalization + cg.num_association + cg.num_aggregation + cg.num_attribute); + double ratio_aggregation = 100.0 * (double) cg.num_aggregation / (double) (cg.num_generalization + cg.num_association + cg.num_aggregation + cg.num_attribute); + double ratio_attribute = 100.0 * (double) cg.num_attribute / (double) (cg.num_generalization + cg.num_association + cg.num_aggregation + cg.num_attribute); + System.out.println(ratio_generalization); + System.out.println(ratio_association); + System.out.println(ratio_aggregation); + System.out.println(ratio_attribute); + } + + private static void read(String doc_dir, CountGroup cg) { + try { + // 工作簿 + Workbook book = Workbook.getWorkbook(new File(doc_dir)); + // 获得第一个工作表对象 + Sheet sheet = book.getSheet("Sheet0"); + int rows = sheet.getRows(); + // 对每行数据,解析关系类型 + for (int row = 1; row < rows; row++) { + String relation_str = sheet.getCell(4, row).getContents().trim(); + switch (relation_str) { + case "GENERALIZATION": { + cg.num_generalization++; + break; + } + case "ASSOCIATION": { + cg.num_association++; + break; + } + case "AGGREGATION": { + cg.num_aggregation++; + break; + } + case "ATTRIBUTE": { + cg.num_attribute++; + break; + } + } + } + book.close(); + } catch (BiffException e) { + e.printStackTrace(); + } catch (IOException e) { + e.printStackTrace(); + } + } + + private static class CountGroup { + int num_generalization = 0; + int num_association = 0; + int num_aggregation = 0; + int num_attribute = 0; + } +} diff --git a/src/main/java/com/hy/java/uct/cdtocode/mapper/CodeRelationMapper.java b/src/main/java/com/hy/java/uct/cdtocode/mapper/CodeRelationMapper.java index bd1a803761c961f7a011c3c86e687cef4ad212c8..21c62168d293e579348d143c5829758c686df9ec 100644 --- a/src/main/java/com/hy/java/uct/cdtocode/mapper/CodeRelationMapper.java +++ b/src/main/java/com/hy/java/uct/cdtocode/mapper/CodeRelationMapper.java @@ -43,20 +43,22 @@ import jxl.write.biff.RowsExceededException; public class CodeRelationMapper { // 文档实体与代码的相似度阈值。小项目0.6左右,大项目0.75左右 - private static final double sim_threshold = 0.75; + private static final double sim_threshold = 0.6; // 同一包下文档实体与代码的相似度阈值。小项目0.25左右,大项目0.65左右 - private static final double sim_p_threshold = 0.65; + private static final double sim_p_threshold = 0.25; // 包名相似度阈值。小项目0.85左右,大项目1.0左右 - private static final double sim_pck_threshold = 1.0; - // 文档实体平均追踪概率阈值。小项目0.35左右,大项目0.7~0.75左右 - private static final double ave_threshold = 0.75; - // 类图最终追踪概率阈值。小项目0.55左右,大项目0.75~0.8左右 - private static final double pre_threshold = 0.8; + private static final double sim_pck_threshold = 0.85; + // 文档实体平均追踪概率阈值。小项目0.35左右,大项目0.75左右 + private static final double ave_threshold = 0.35; + // 类图最终追踪概率阈值。小项目0.55左右,大项目0.8左右 + private static final double pre_threshold = 0.55; /** * 对扩展后的模型与代码进行匹配 + * + * @param b */ - public static Map map(Map classShortName_classObj_mappedByDoc, Map classFullName_javaFileDir_map) { + public static Map map(Map classShortName_classObj_mappedByDoc, Map classFullName_javaFileDir_map, boolean _do) { System.out.println("开始对扩展后的模型与代码进行匹配"); Set ClsImg_shortName_set = classShortName_classObj_mappedByDoc.keySet(); /* @@ -99,7 +101,7 @@ public class CodeRelationMapper { * 进行完这步后,对Cls_img的每个Ent_doc而言,会将其追踪到数个truly_mapped_file,每个truly_mapped_file均对应数个PE_code(Ent_doc,Cls_code,R) */ System.out.println("开始基于关系推理,对模糊匹配的追踪结果进行取舍"); - detectUMLCLsRelations(classShortName_classObj_mappedByDoc, classFullName_javaFileDir_map); + detectUMLCLsRelations(classShortName_classObj_mappedByDoc, classFullName_javaFileDir_map, _do); System.out.println("完成基于关系推理,对模糊匹配的追踪结果进行取舍"); // 检查PE_code(Ent_doc,Cls_code,R),可注释掉 // check_PEcodeR(classShortName_classObj_mappedByDoc); @@ -376,15 +378,63 @@ public class CodeRelationMapper { * 对每个Ent_doc,查其所有追踪到的Cls_code是否与Ent_doc.related_Ent有关系。如果一个Cls_code与某个由related_Ent追踪到的Cls_code'有同类关系,则该Cls_code就是“根据关系推理追踪到的代码文件”。 * * 保留所有由推理获得的“可能追踪到的类”(可以新存到一个list中),舍弃所有不能从关系推理到的类 + * + * @param b */ - private static void detectUMLCLsRelations(Map classShortName_classObj_mappedByDoc, Map classFullName_javaFileDir_map) { + private static void detectUMLCLsRelations(Map classShortName_classObj_mappedByDoc, Map classFullName_javaFileDir_map, boolean _do) { Set ClsImg_shortName_set = classShortName_classObj_mappedByDoc.keySet(); for (String ClsImg_shortName : ClsImg_shortName_set) { UMLClass UML_class = classShortName_classObj_mappedByDoc.get(ClsImg_shortName); // 对每个Ent_doc,查其所有追踪到的Cls_code是否与Ent_doc.related_Ent有关系 for (Entity Ent_doc : UML_class.mappedEnt_ls) { for (Pair possibleMapped_javaFile_pair : Ent_doc.possibleMapped_javaFiles) { - if (possibleMapped_javaFile_pair.getRight() == 1.0) { + if (_do) { + if (possibleMapped_javaFile_pair.getRight() == 1.0) { + // 记录这个“根据关系推理追踪到的代码文件”。 + MappedFile truly_mapped_file_forEnt = new MappedFile(); + truly_mapped_file_forEnt.java_file_dir = classFullName_javaFileDir_map.get(possibleMapped_javaFile_pair.getLeft()); + // PE_code(Ent_doc,Cls_code,R) + // 如果Ent_code能纯粹根据文本相似度追踪到Cls_code,则将PEcode_R中的PR_doc部分置为默认值1 + double PEcode_R = 0.5 * possibleMapped_javaFile_pair.getRight() + 0.5 * 1.0; + truly_mapped_file_forEnt.PEcode_R_ls.add(Pair.createPair(null, PEcode_R)); + Ent_doc.trulyMapped_javaFiles.add(truly_mapped_file_forEnt); + } else { + try { + CompilationUnit possibleMapped_ClsCode = StaticJavaParser.parse(new File(classFullName_javaFileDir_map.get(possibleMapped_javaFile_pair.getLeft()))); + String cls_shortName = getClsShortNameFromFullName(possibleMapped_javaFile_pair.getLeft()); + /* + * 如果一个Cls_code与某个由related_Ent追踪到的Cls_code'有同类关系,则该Cls_code就是“根据关系推理追踪到的代码文件”。 + * + * 注意: 如果Ent_code没有related_Ent,则直接认为Ent_code能纯粹根据文本相似度追踪到Cls_code + */ + Pair> ifTrulyMapped_possibleRs = check_if_trulyMapped(possibleMapped_ClsCode, cls_shortName, possibleMapped_javaFile_pair.getLeft(), Ent_doc.relations_between_Ent); + // 如果确实真的可以这么追踪,则记录这个文件,并计算追踪概率 + if (ifTrulyMapped_possibleRs.getLeft() == true) { + // 记录这个“根据关系推理追踪到的代码文件”。 + MappedFile truly_mapped_file_forEnt = new MappedFile(); + truly_mapped_file_forEnt.java_file_dir = classFullName_javaFileDir_map.get(possibleMapped_javaFile_pair.getLeft()); + // 记录在每条关系R下从Ent_doc追踪到Cls_code的概率PEcode_R + if (ifTrulyMapped_possibleRs.getRight().size() <= 0) { + // PE_code(Ent_doc,Cls_code,R) + // 如果Ent_code能纯粹根据文本相似度追踪到Cls_code,则将PEcode_R中的PR_doc部分置为默认值1 + double PEcode_R = 0.5 * possibleMapped_javaFile_pair.getRight() + 0.5 * 1.0; + truly_mapped_file_forEnt.PEcode_R_ls.add(Pair.createPair(null, PEcode_R)); + } else { + // 记录在每条关系R下从Ent_doc追踪到Cls_code的概率PEcode_R + for (EntityRelation R : ifTrulyMapped_possibleRs.getRight()) { + // PE_code(Ent_doc,Cls_code,R) + double PEcode_R = 0.5 * possibleMapped_javaFile_pair.getRight() + 0.5 * R.PR_doc; + truly_mapped_file_forEnt.PEcode_R_ls.add(Pair.createPair(R, PEcode_R)); + } + } + Ent_doc.trulyMapped_javaFiles.add(truly_mapped_file_forEnt); + // 进行到这儿后,对每个Ent_doc而言,会将其追踪到数个truly_mapped_file,每个truly_mapped_file均对应数个PE_code(Ent_doc,Cls_code,R) + } + } catch (FileNotFoundException e) { + e.printStackTrace(); + } + } + } else { // 记录这个“根据关系推理追踪到的代码文件”。 MappedFile truly_mapped_file_forEnt = new MappedFile(); truly_mapped_file_forEnt.java_file_dir = classFullName_javaFileDir_map.get(possibleMapped_javaFile_pair.getLeft()); @@ -393,41 +443,6 @@ public class CodeRelationMapper { double PEcode_R = 0.5 * possibleMapped_javaFile_pair.getRight() + 0.5 * 1.0; truly_mapped_file_forEnt.PEcode_R_ls.add(Pair.createPair(null, PEcode_R)); Ent_doc.trulyMapped_javaFiles.add(truly_mapped_file_forEnt); - } else { - try { - CompilationUnit possibleMapped_ClsCode = StaticJavaParser.parse(new File(classFullName_javaFileDir_map.get(possibleMapped_javaFile_pair.getLeft()))); - String cls_shortName = getClsShortNameFromFullName(possibleMapped_javaFile_pair.getLeft()); - /* - * 如果一个Cls_code与某个由related_Ent追踪到的Cls_code'有同类关系,则该Cls_code就是“根据关系推理追踪到的代码文件”。 - * - * 注意: 如果Ent_code没有related_Ent,则直接认为Ent_code能纯粹根据文本相似度追踪到Cls_code - */ - Pair> ifTrulyMapped_possibleRs = check_if_trulyMapped(possibleMapped_ClsCode, cls_shortName, possibleMapped_javaFile_pair.getLeft(), Ent_doc.relations_between_Ent); - // 如果确实真的可以这么追踪,则记录这个文件,并计算追踪概率 - if (ifTrulyMapped_possibleRs.getLeft() == true) { - // 记录这个“根据关系推理追踪到的代码文件”。 - MappedFile truly_mapped_file_forEnt = new MappedFile(); - truly_mapped_file_forEnt.java_file_dir = classFullName_javaFileDir_map.get(possibleMapped_javaFile_pair.getLeft()); - // 记录在每条关系R下从Ent_doc追踪到Cls_code的概率PEcode_R - if (ifTrulyMapped_possibleRs.getRight().size() <= 0) { - // PE_code(Ent_doc,Cls_code,R) - // 如果Ent_code能纯粹根据文本相似度追踪到Cls_code,则将PEcode_R中的PR_doc部分置为默认值1 - double PEcode_R = 0.5 * possibleMapped_javaFile_pair.getRight() + 0.5 * 1.0; - truly_mapped_file_forEnt.PEcode_R_ls.add(Pair.createPair(null, PEcode_R)); - } else { - // 记录在每条关系R下从Ent_doc追踪到Cls_code的概率PEcode_R - for (EntityRelation R : ifTrulyMapped_possibleRs.getRight()) { - // PE_code(Ent_doc,Cls_code,R) - double PEcode_R = 0.5 * possibleMapped_javaFile_pair.getRight() + 0.5 * R.PR_doc; - truly_mapped_file_forEnt.PEcode_R_ls.add(Pair.createPair(R, PEcode_R)); - } - } - Ent_doc.trulyMapped_javaFiles.add(truly_mapped_file_forEnt); - // 进行到这儿后,对每个Ent_doc而言,会将其追踪到数个truly_mapped_file,每个truly_mapped_file均对应数个PE_code(Ent_doc,Cls_code,R) - } - } catch (FileNotFoundException e) { - e.printStackTrace(); - } } } } diff --git a/src/main/java/com/hy/java/uct/cdtocode/vsmlsi/SimNameTracer.java b/src/main/java/com/hy/java/uct/cdtocode/vsmlsi/SimNameTracer.java new file mode 100644 index 0000000000000000000000000000000000000000..3d768b1fa61a0b1c00288578c1b6696a12f22ef6 --- /dev/null +++ b/src/main/java/com/hy/java/uct/cdtocode/vsmlsi/SimNameTracer.java @@ -0,0 +1,107 @@ +package com.hy.java.uct.cdtocode.vsmlsi; + +import java.util.Collection; +import java.util.HashMap; +import java.util.Map; + +import com.hy.java.uct.cdtocode.reader.CDReader; +import com.hy.java.uct.cdtocode.reader.CodeReader; +import com.hy.java.uct.util.UMLClass; + +public class SimNameTracer { + /** + * 将要追踪的类图放在cd_dir目录下 + */ + private static final String cd_dir = System.getProperty("user.dir") + "\\src\\main\\resources\\cdtocode\\cd\\"; + + /** + * 将要追踪的代码放在code_dir目录下 + */ + private static final String code_dir = System.getProperty("user.dir") + "\\src\\main\\resources\\cdtocode\\code\\"; + + /** + * 将追踪结果放在res_dir目录下 + */ + private static final String res_dir = System.getProperty("user.dir") + "\\src\\main\\resources\\cdtocode\\"; + + /** + * 针对类图及其相关文档,追踪到代码中的类 + */ + public static void main(String[] args) { + /* + * 1、读取模型信息 + */ + // 读取完UML图识别结果后,将实体信息保存在classes_in_CD里。形式为 + /* + * Apache OODT File Manager + */ + // Map classes_in_CD = CDReader.read(cd_dir + "cd-Apache OODT File Manager.txt"); + /* + * Hadoop HDFS + */ + // Map classes_in_CD = CDReader.read(cd_dir + "cd-Hadoop HDFS.txt"); + /* + * Hadoop MapReduce + */ + Map classes_in_CD = CDReader.read(cd_dir + "cd-Hadoop MapReduce.txt"); + // 检查结果,可注释掉 + // CDReader.check(classes_in_CD); + /* + * 2、读取code path指定的目录下所有java文件 + * + * <类全称(包+类名), java_file_path> + */ + // 记得改这里的路径 + // Map classFullName_javaFileDir_map = CodeReader.read(code_dir + "code path-fm"); + // Map classFullName_javaFileDir_map = CodeReader.read(code_dir + "code path-hdfs"); + Map classFullName_javaFileDir_map = CodeReader.read(code_dir + "code path-mr"); + // 检查结果,可注释掉 + // CodeReader.check(classFullName_javaFileDir_map); + // 针对每个图中的类名,如果代码中有相似的包名,则将包下所有相似的类(字符串包含至少一个关键词即可)都重点标记。此脚本为一次性,使用后删除即可 + // Set packs = CodeReader.findPcks(code_dir + "code path-fm"); + // Set packs = CodeReader.findPcks(code_dir + "code path-hdfs"); + // Set packs = CodeReader.findPcks(code_dir + "code path-mr"); + /* + * 3、遍历模型中的实体元素(类或对象),针对每个元素,在code中寻找能匹配的java文件 + * + * 基于启发式模糊匹配的UML类图与代码追踪方法:首先针对类图中的类和代码中的类,基于类的名称进行字符串完全匹配,从而建立确定的初始追踪;基于同义词、 词缀词典等语料库,定义基于类名称匹配和关联关系的启发式匹配规则,研究基于模糊匹配技术的追踪关系建立方法,基于初始追踪和启发式规则, 对类名不一致的模型和代码元素进行启发式追踪,扩展初始追踪关系。 + */ + Map mapped_classes = SimNameTracer.map(classes_in_CD, classFullName_javaFileDir_map); + /* + * Apache OODT File Manager + */ + // SimNameTracer.save(mapped_classes, res_dir + "Apache OODT File Manager-simname.xls"); + /* + * Hadoop HDFS + */ + // SimNameTracer.save(mapped_classes, res_dir + "Hadoop HDFS-simname.xls"); + /* + * Hadoop MapReduce + */ + SimNameTracer.save(mapped_classes, res_dir + "Hadoop MapReduce-simname.xls"); + // 检查结果,可注释掉 + // SimNameTracer.check(res_dir + "Apache OODT File Manager.xls"); + } + + private static Map map(Map classes_in_CD, Map classFullName_javaFileDir_map) { + Map res = new HashMap<>(); + Collection classes = classes_in_CD.values(); + // 对每个类,根据其类名去匹配代码 + for (UMLClass cls_img : classes) { + + cls_img.mappedJavaFiles.add(null); + } + return res; + } + + /** + * 通过类全称,获得类短称 + */ + private static String getClsShortNameFromFullName(String clsCode_fullName) { + return clsCode_fullName.substring(clsCode_fullName.lastIndexOf(".") + 1); + } + + private static void save(Map mapped_classes, String res_dir) { + // TODO Auto-generated method stub + } +} diff --git a/src/main/java/com/hy/java/uct/cdtocode/vsmlsi/VSMAndLSIDataGenerator.java b/src/main/java/com/hy/java/uct/cdtocode/vsmlsi/VSMAndLSIDataGenerator.java index de71004947d94b3977ec2d6ed97f7acc51a300f7..00d5190b0cfeec241088a25f9fc4e17545970d0d 100644 --- a/src/main/java/com/hy/java/uct/cdtocode/vsmlsi/VSMAndLSIDataGenerator.java +++ b/src/main/java/com/hy/java/uct/cdtocode/vsmlsi/VSMAndLSIDataGenerator.java @@ -1,27 +1,14 @@ package com.hy.java.uct.cdtocode.vsmlsi; import java.io.File; -import java.io.FileNotFoundException; import java.io.IOException; -import java.util.ArrayList; -import java.util.List; import java.util.Map; -import java.util.Optional; import java.util.Set; -import com.github.javaparser.StaticJavaParser; -import com.github.javaparser.ast.CompilationUnit; -import com.github.javaparser.ast.NodeList; -import com.github.javaparser.ast.body.ClassOrInterfaceDeclaration; -import com.github.javaparser.ast.body.FieldDeclaration; -import com.github.javaparser.ast.body.MethodDeclaration; -import com.github.javaparser.ast.type.ClassOrInterfaceType; import com.hy.java.uct.cdtocode.reader.CDReader; import com.hy.java.uct.cdtocode.reader.CodeReader; -import com.hy.java.uct.cdtocode.reader.DocReader; import com.hy.java.uct.util.UMLClass; import com.hy.java.utility.common.FileEditor; -import com.hy.java.utility.common.Pair; import jxl.Sheet; import jxl.Workbook; @@ -38,11 +25,6 @@ public class VSMAndLSIDataGenerator { */ private static final String cd_dir = System.getProperty("user.dir") + "\\src\\main\\resources\\cdtocode\\cd\\"; - /** - * 将设计文档放在doc_dir目录下 - */ - private static final String doc_dir = System.getProperty("user.dir") + "\\src\\main\\resources\\cdtocode\\doc\\"; - /** * 将要追踪的代码放在code_dir目录下 */ @@ -61,15 +43,15 @@ public class VSMAndLSIDataGenerator { /* * Apache OODT File Manager */ - Map classes_in_CD = CDReader.read(cd_dir + "cd-Apache OODT File Manager.txt"); + Map classes_in_CD_fm = CDReader.read(cd_dir + "cd-Apache OODT File Manager.txt"); /* * Hadoop HDFS */ - // Map classes_in_CD = CDReader.read(cd_dir + "cd-Hadoop HDFS.txt"); + Map classes_in_CD_hdfs = CDReader.read(cd_dir + "cd-Hadoop HDFS.txt"); /* * Hadoop MapReduce */ - // Map classes_in_CD = CDReader.read(cd_dir + "cd-Hadoop MapReduce.txt"); + Map classes_in_CD_mr = CDReader.read(cd_dir + "cd-Hadoop MapReduce.txt"); // 检查结果,可注释掉 // CDReader.check(classes_in_CD); /* @@ -78,9 +60,9 @@ public class VSMAndLSIDataGenerator { * <类全称(包+类名), java_file_path> */ // 记得改这里的路径 - Map classFullName_javaFileDir_map = CodeReader.read(code_dir + "code path-fm"); - // Map classFullName_javaFileDir_map = CodeReader.read(code_dir + "code path-hdfs"); - // Map classFullName_javaFileDir_map = CodeReader.read(code_dir + "code path-mr"); + Map classFullName_javaFileDir_map_fm = CodeReader.read(code_dir + "code path-fm"); + Map classFullName_javaFileDir_map_hdfs = CodeReader.read(code_dir + "code path-hdfs"); + Map classFullName_javaFileDir_map_mr = CodeReader.read(code_dir + "code path-mr"); // 检查结果,可注释掉 // CodeReader.check(classFullName_javaFileDir_map); /* @@ -88,15 +70,15 @@ public class VSMAndLSIDataGenerator { * * 生成一个excel文件,每一行包含的是:类名 属性 方法 */ - VSMAndLSIDataGenerator.save(classes_in_CD, classFullName_javaFileDir_map, res_dir + "Apache OODT File Manager-VSMLSIdata.xls"); - // VSMAndLSIDataGenerator.save(classes_in_CD, classFullName_javaFileDir_map, res_dir + "Hadoop HDFS-VSMLSIdata.xls"); - // VSMAndLSIDataGenerator.save(classes_in_CD, classFullName_javaFileDir_map, res_dir + "Hadoop MapReduce-VSMLSIdata.xls"); + VSMAndLSIDataGenerator.save(classes_in_CD_fm, classFullName_javaFileDir_map_fm, res_dir + "Apache OODT File Manager-VSMLSIdata.xls"); + VSMAndLSIDataGenerator.save(classes_in_CD_hdfs, classFullName_javaFileDir_map_hdfs, res_dir + "Hadoop HDFS-VSMLSIdata.xls"); + VSMAndLSIDataGenerator.save(classes_in_CD_mr, classFullName_javaFileDir_map_mr, res_dir + "Hadoop MapReduce-VSMLSIdata.xls"); /* * 4、读取excel文件,生成可用数据 */ VSMAndLSIDataGenerator.generate(res_dir + "Apache OODT File Manager-VSMLSIdata.xls", res_dir + "Apache OODT File Manager-VSMLSI.txt"); - // VSMAndLSIDataGenerator.generate(res_dir + "Hadoop HDFS-VSMLSIdata.xls", res_dir + "Hadoop HDFS-VSMLSI.txt"); - // VSMAndLSIDataGenerator.generate(res_dir + "Hadoop MapReduce-VSMLSIdata.xls", res_dir + "Hadoop MapReduce-VSMLSI.txt"); + VSMAndLSIDataGenerator.generate(res_dir + "Hadoop HDFS-VSMLSIdata.xls", res_dir + "Hadoop HDFS-VSMLSI.txt"); + VSMAndLSIDataGenerator.generate(res_dir + "Hadoop MapReduce-VSMLSIdata.xls", res_dir + "Hadoop MapReduce-VSMLSI.txt"); } /** @@ -148,6 +130,7 @@ public class VSMAndLSIDataGenerator { private static int saveUMLCls(WritableSheet sheet, int row, UMLClass UML_class) { try { + // sheet.addCell(new Label(0, row, UML_class.getTitle())); sheet.addCell(new Label(0, row, UML_class.getTitle() + " " + UML_class.getAttrisStr().replaceAll("\n", " ") + " " + UML_class.getMethodsStr().replaceAll("\n", " "))); row++; } catch (RowsExceededException e) { @@ -165,41 +148,19 @@ public class VSMAndLSIDataGenerator { String temp = ClsCode_fullName.substring(0, ClsCode_fullName.lastIndexOf(".")); String pck = temp.substring(temp.lastIndexOf(".") + 1); String cls_shortName = getClsShortNameFromFullName(ClsCode_fullName); - content += pck + "." + cls_shortName; - try { - CompilationUnit clsCode = StaticJavaParser.parse(new File(cls_file_dir)); - Optional o_class = clsCode.getClassByName(cls_shortName); - if (o_class.isPresent()) { - ClassOrInterfaceDeclaration clsCode_unit = o_class.get(); - /* - * 先比较parent_cls的所有属性 - */ - List ClsCode_fields = clsCode_unit.getFields(); - // 对比UMLAttris和ClsCode_fields - for (FieldDeclaration attri : ClsCode_fields) { - content += " " + attri.getVariable(0).getNameAsString(); - } - /* - * 再比较parent_cls的所有方法 - */ - List ClsCode_methods = clsCode_unit.getMethods(); - // 对比UMLMethods和ClsCode_methods - for (MethodDeclaration method : ClsCode_methods) { - content += " " + method.getNameAsString(); - } - } - } catch (FileNotFoundException e) { - // TODO Auto-generated catch block - e.printStackTrace(); - } + content += pck + " " + cls_shortName; + /* + * try { CompilationUnit clsCode = StaticJavaParser.parse(new File(cls_file_dir)); Optional o_class = clsCode.getClassByName(cls_shortName); if (o_class.isPresent()) { + * ClassOrInterfaceDeclaration clsCode_unit = o_class.get(); List ClsCode_fields = clsCode_unit.getFields(); for (FieldDeclaration attri : ClsCode_fields) { content += " " + + * attri.getVariable(0).getNameAsString(); } List ClsCode_methods = clsCode_unit.getMethods(); for (MethodDeclaration method : ClsCode_methods) { content += " " + + * method.getNameAsString(); } } } catch (FileNotFoundException e) { e.printStackTrace(); } + */ try { sheet.addCell(new Label(0, row, content)); row++; } catch (RowsExceededException e) { - // TODO Auto-generated catch block e.printStackTrace(); } catch (WriteException e) { - // TODO Auto-generated catch block e.printStackTrace(); } return row; @@ -223,28 +184,18 @@ public class VSMAndLSIDataGenerator { // Sheet sheet = book.getSheet(0); int rows = sheet.getRows(); for (int row = 0; row < rows; row++) { - String class_info = sheet.getCell(0, row).getContents(); + String class_info = sheet.getCell(0, row).getContents().trim(); // 根据python脚本格式编写内容 - // 给我好好写!!! - // 给我好好写!!! - // 给我好好写!!! - // 给我好好写!!! - // 给我好好写!!! - // 给我好好写!!! - // 给我好好写!!! - // 给我好好写!!! - // 给我好好写!!! - // 给我好好写!!! - // 给我好好写!!! - // 给我好好写!!! - fe.write(class_info, true); + if (row == 0) { + fe.write("\"" + class_info + "\"", true); + } else { + fe.write(",\n" + "\"" + class_info + "\"", true); + } } book.close(); } catch (BiffException e) { - // TODO Auto-generated catch block e.printStackTrace(); } catch (IOException e) { - // TODO Auto-generated catch block e.printStackTrace(); } } diff --git a/src/main/java/com/hy/java/uct/cdtocode/vsmlsi/VSMAndLSIParser.java b/src/main/java/com/hy/java/uct/cdtocode/vsmlsi/VSMAndLSIParser.java index 265744a70ac3c6004b0021263ad4d8c3f7e6bf2f..6a2af08a0a388ac21f05fc523b10077c4c236578 100644 --- a/src/main/java/com/hy/java/uct/cdtocode/vsmlsi/VSMAndLSIParser.java +++ b/src/main/java/com/hy/java/uct/cdtocode/vsmlsi/VSMAndLSIParser.java @@ -1,12 +1,125 @@ package com.hy.java.uct.cdtocode.vsmlsi; +import java.io.File; +import java.io.IOException; +import java.util.Map; +import java.util.Set; + +import com.hy.java.uct.cdtocode.reader.CDReader; +import com.hy.java.uct.util.UMLClass; +import com.hy.java.utility.common.FileEditor; + +import jxl.Sheet; +import jxl.Workbook; +import jxl.read.biff.BiffException; +import jxl.write.Label; +import jxl.write.WritableSheet; +import jxl.write.WritableWorkbook; +import jxl.write.WriteException; +import jxl.write.biff.RowsExceededException; + /** * 解析VSM和LSI结果文件(csv文件) */ public class VSMAndLSIParser { + /** + * 将要追踪的类图放在cd_dir目录下 + */ + private static final String cd_dir = System.getProperty("user.dir") + "\\src\\main\\resources\\cdtocode\\cd\\"; + + /** + * 将追踪结果放在res_dir目录下 + */ + private static final String res_dir = System.getProperty("user.dir") + "\\src\\main\\resources\\cdtocode\\"; public static void main(String[] args) { - // TODO Auto-generated method stub + /* + * 1、读取模型信息 + */ + // 读取完UML图识别结果后,将实体信息保存在classes_in_CD里。形式为 + /* + * Apache OODT File Manager + */ + Map classes_in_CD_fm = CDReader.read(cd_dir + "cd-Apache OODT File Manager.txt"); + /* + * Hadoop HDFS + */ + Map classes_in_CD_hdfs = CDReader.read(cd_dir + "cd-Hadoop HDFS.txt"); + /* + * Hadoop MapReduce + */ + Map classes_in_CD_mr = CDReader.read(cd_dir + "cd-Hadoop MapReduce.txt"); + // 检查结果,可注释掉 + // CDReader.check(classes_in_CD); + /* + * 2、读取excel文件,生成可用数据 + */ + VSMAndLSIParser.analyze(classes_in_CD_fm, res_dir + "Apache OODT File Manager-VSMLSIdata.xls", res_dir + "Apache OODT File Manager-vsmlsiraw.xls", res_dir + "Apache OODT File Manager-VSMLSI.xls"); + VSMAndLSIParser.analyze(classes_in_CD_hdfs, res_dir + "Hadoop HDFS-VSMLSIdata.xls", res_dir + "Hadoop HDFS-vsmlsiraw.xls", res_dir + "Hadoop HDFS-VSMLSI.xls"); + VSMAndLSIParser.analyze(classes_in_CD_mr, res_dir + "Hadoop MapReduce-VSMLSIdata.xls", res_dir + "Hadoop MapReduce-vsmlsiraw.xls", res_dir + "Hadoop MapReduce-VSMLSI.xls"); + } + private static void analyze(Map classes_in_CD, String data_dir, String raw_dir, String res_dir) { + int num_of_cls = classes_in_CD.size(); + try { + // 最终结果工作簿 + WritableWorkbook res_wb = Workbook.createWorkbook(new File(res_dir)); + // 新建工作表 + WritableSheet res_sheet_vsm = res_wb.createSheet("vsm", 0); + WritableSheet res_sheet_lsi = res_wb.createSheet("lsi", 1); + // 数据列表工作簿 + Workbook data_wb = Workbook.getWorkbook(new File(data_dir)); + Sheet data_sheet = data_wb.getSheet("Sheet1"); + // 原始结果工作簿 + Workbook rawres_wb = Workbook.getWorkbook(new File(raw_dir)); + Sheet rawres_sheet_vsm = rawres_wb.getSheet("vsm"); + Sheet rawres_sheet_lsi = rawres_wb.getSheet("lsi"); + // 读原始结果(借助数据工作簿),写最终结果 + // vsm + int rows_vsm = rawres_sheet_vsm.getRows(); + // 针对每个类,遍历所有结果,选取概率较大的 + for (int col = 1; col < num_of_cls + 1; col++) { + // 借助数据工作簿,获得当前类的类名,并添加到结果的表头 + String cls_name = data_sheet.getCell(0, col - 1).getContents().trim(); + res_sheet_vsm.addCell(new Label(col, 0, cls_name)); + for (int row = num_of_cls + 1; row < rows_vsm; row++) { + String code_name = data_sheet.getCell(0, row - 1).getContents().trim(); + res_sheet_vsm.addCell(new Label(0, row - num_of_cls, code_name)); + String trace_info = rawres_sheet_vsm.getCell(col, row).getContents().trim(); + res_sheet_vsm.addCell(new Label(col, row - num_of_cls, trace_info)); + } + } + // lsi + int rows_lsi = rawres_sheet_lsi.getRows(); + // 针对每个类,遍历所有结果,选取概率较大的 + for (int col = 1; col < num_of_cls + 1; col++) { + // 借助数据工作簿,获得当前类的类名,并添加到结果的表头 + String cls_name = data_sheet.getCell(0, col - 1).getContents().trim(); + res_sheet_lsi.addCell(new Label(col, 0, cls_name)); + for (int row = num_of_cls + 1; row < rows_lsi; row++) { + String code_name = data_sheet.getCell(0, row - 1).getContents().trim(); + res_sheet_lsi.addCell(new Label(0, row - num_of_cls, code_name)); + String trace_info = rawres_sheet_lsi.getCell(col, row).getContents().trim(); + res_sheet_lsi.addCell(new Label(col, row - num_of_cls, trace_info)); + } + } + // 写入文件 + res_wb.write(); + res_wb.close(); + data_wb.close(); + rawres_wb.close(); + } catch (RowsExceededException e) { + // TODO Auto-generated catch block + e.printStackTrace(); + } catch (BiffException e) { + // TODO Auto-generated catch block + e.printStackTrace(); + } catch (WriteException e) { + // TODO Auto-generated catch block + e.printStackTrace(); + } catch (IOException e) { + // TODO Auto-generated catch block + e.printStackTrace(); + } } } diff --git a/src/main/java/com/hy/java/uct/umlrecog/UMLDiagramRecognizer.java b/src/main/java/com/hy/java/uct/umlrecog/UMLDiagramRecognizer.java index f52c7a532f1632ee0aebf2c2effcb331dd2e3454..30b57e98d4913e0def1377ff1afc08b9e66086a7 100644 --- a/src/main/java/com/hy/java/uct/umlrecog/UMLDiagramRecognizer.java +++ b/src/main/java/com/hy/java/uct/umlrecog/UMLDiagramRecognizer.java @@ -46,6 +46,7 @@ public class UMLDiagramRecognizer { */ public static void main(String[] args) { // 类图实验 + // true的是要识别的 UMLDiagramRecognizer.recogCD(cd_dir, "abrden/StarCraft", false); UMLDiagramRecognizer.recogCD(cd_dir, "alexasahis/km2", false); UMLDiagramRecognizer.recogCD(cd_dir, "BackupTheBerlios/jpwgen-svn", false); @@ -56,14 +57,13 @@ public class UMLDiagramRecognizer { UMLDiagramRecognizer.recogCD(cd_dir, "dsarlis/SoftEng", false); UMLDiagramRecognizer.recogCD(cd_dir, "egoless/pqs", false); UMLDiagramRecognizer.recogCD(cd_dir, "emeric254/Java-STRI-S4", false); - UMLDiagramRecognizer.recogCD(cd_dir, "bloodmarry12/rapid-framework_googlecode", false); - UMLDiagramRecognizer.recogCD(cd_dir, "rodrigoazevedomartins/TrabalhoFinalLTPIII", false); + UMLDiagramRecognizer.recogCD(cd_dir, "rodrigoazevedomartins/TrabalhoFinalLTPIII", true);// UMLDiagramRecognizer.recogCD(cd_dir, "ultragdb/org.eclipse.cdt", false); UMLDiagramRecognizer.recogCD(cd_dir, "georgejakes/Xug", false); UMLDiagramRecognizer.recogCD(cd_dir, "hungnguyen94/BTrouble", false); UMLDiagramRecognizer.recogCD(cd_dir, "klemens/openolat", false); - UMLDiagramRecognizer.recogCD(cd_dir, "lorneliechty/pleaseholdapplause", false); + UMLDiagramRecognizer.recogCD(cd_dir, "lorneliechty/pleaseholdapplause", true);// UMLDiagramRecognizer.recogCD(cd_dir, "neuroph/neuroph", false); UMLDiagramRecognizer.recogCD(cd_dir, "PillowSoPaw/SPSWENG-astroNATS", false); UMLDiagramRecognizer.recogCD(cd_dir, "tiendan3108/CP", false); @@ -72,7 +72,7 @@ public class UMLDiagramRecognizer { UMLDiagramRecognizer.recogCD(cd_dir, "fltt/jss7", false); UMLDiagramRecognizer.recogCD(cd_dir, "gemxd/gemfirexd-oss", false); UMLDiagramRecognizer.recogCD(cd_dir, "ging/isabel", false); - UMLDiagramRecognizer.recogCD(cd_dir, "hangum/TadpoleForDBTools", false); + UMLDiagramRecognizer.recogCD(cd_dir, "hangum/TadpoleForDBTools", true);// UMLDiagramRecognizer.recogCD(cd_dir, "Istarnion/Team12", false); UMLDiagramRecognizer.recogCD(cd_dir, "jaissegrela/steganography", false); UMLDiagramRecognizer.recogCD(cd_dir, "jguze/ACiv", false); @@ -89,13 +89,13 @@ public class UMLDiagramRecognizer { UMLDiagramRecognizer.recogCD(cd_dir, "retoo/bodesuri", false); UMLDiagramRecognizer.recogCD(cd_dir, "richardimms/PRCSA", false); UMLDiagramRecognizer.recogCD(cd_dir, "rNdm74/Java", false); - UMLDiagramRecognizer.recogCD(cd_dir, "Salaboy/smart-tasks", false); + UMLDiagramRecognizer.recogCD(cd_dir, "Salaboy/smart-tasks", true);// UMLDiagramRecognizer.recogCD(cd_dir, "Sanchez82/Crawler", false); - UMLDiagramRecognizer.recogCD(cd_dir, "snucsne/bio-inspired-leadership", false); + UMLDiagramRecognizer.recogCD(cd_dir, "snucsne/bio-inspired-leadership", true);// UMLDiagramRecognizer.recogCD(cd_dir, "steyskal/Ren-Fest", false); UMLDiagramRecognizer.recogCD(cd_dir, "Team-MWSU/GroupProject", false); UMLDiagramRecognizer.recogCD(cd_dir, "teopalva/travel-dream", false); - UMLDiagramRecognizer.recogCD(cd_dir, "valentinamata/flow3in2013", false); + UMLDiagramRecognizer.recogCD(cd_dir, "valentinamata/flow3in2013", true);// UMLDiagramRecognizer.recogCD(cd_dir, "WiReSEP/Rollercoaster2011", false); // 顺序图实验 UMLDiagramRecognizer.recogSD(sd_dir, "laurikin/java-tetris", false); diff --git a/src/main/resources/cdtocode/Apache OODT File Manager-VSMLSI.txt b/src/main/resources/cdtocode/Apache OODT File Manager-VSMLSI.txt index 700433cd222be4e07037b99e65ac5bf5286ce4bd..2b21311fd0a0659732251b186d92e7b171f6c736 100644 --- a/src/main/resources/cdtocode/Apache OODT File Manager-VSMLSI.txt +++ b/src/main/resources/cdtocode/Apache OODT File Manager-VSMLSI.txt @@ -1 +1,194 @@ -Reference Versioner Element Metadata ProductType Product exceptions.CacheException serialVersionUIDexamples.MimeTypeExtractor doConfigure doExtractstructs.Product productId productName productType productStructure references transferStatus rootRef STATUS_TRANSFER STATUS_RECEIVED STRUCTURE_FLAT STRUCTURE_HIERARCHICAL STRUCTURE_STREAM VALID_STRUCTURES LOG productReceivedTime getProductType setProductType getProductStructure setProductStructure getProductReferences setProductReferences getProductName setProductName getProductId setProductId getTransferStatus setTransferStatus getRootRef setRootRef getDefaultFlatProduct toXML parse setProductRecievedTime getProductReceivedTime toStringexceptions.IngestException serialVersionUID maindatatransfer.S3DataTransferer s3Client bucketName encrypt setFileManagerUrl transferProduct retrieveProduct deleteProduct stageFile stripProtocolaction.GetProductByIdCliAction productId getProduct setProductIdquery.QueryResult product metadata toStringFormat getProduct setProduct getMetadata setMetadata getToStringFormat setToStringFormat toString convertMetadataToString concatMetadataIntoString equals hashCodecatalog.LuceneCatalog LOG logger indexDir reader indexFilePath valLayer CATALOG_CACHE generator pageSize writeLockTimeout commitLockTimeout mergeFactor addMetadata removeMetadata addProduct modifyProduct removeProduct setProductTransferStatus addProductReferences getProductById getProductById getCompleteProductById getCompleteProductById getCompleteProductById getProductByName getProductByName getProductReferences getProducts getProducts getProductsByProductType getProductsByProductType getMetadata getReducedMetadata query getTopNProducts getTopNProducts getValidationLayer getNumProducts getFirstPage getLastProductPage getNextPage getPrevPage pagedQuery removeProductDocument addCompleteProductToIndex toCompleteProduct toCompleteProduct toDoc hasMetadataAndRefs getNumHits paginateQuery getQuery quietGetElementsdatatransfer.RemoteDataTransferer NUM_BYTES fileManagerUrl chunkSize client LOG setFileManagerUrl transferProduct retrieveProduct deleteProduct remoteTransfer quietNotifyTransferProduct quietNotifyProductTransferCompletecatalog.DataSourceCatalogFactory VAL VAL1 dataSource validationLayer fieldIdStr pageSize cacheUpdateMinutes lenientFields productIdString orderedValues configure createCatalogrpc.FileManagerServerFactoryutil.XmlStructFactory LOG getProductType writeProductTypeMapXmLDocument writeElementXmlDocument writeProductTypeXmlDocument getProductTypeMapXmlDocument getElementXmlDocument getProductTypeXmlDocument getElement getProductTypeElementList friendlyXmldatatransfer.InPlaceDataTransferer LOG client setFileManagerUrl transferProduct retrieveProduct deleteProductexamples.FinalFileLocationExtractor replaceLocation doConfigure doExtract scrubRefsutil.XmlRpcStructFactory getXmlRpcFileTransferStatus getFileTransferStatusFromXmlRpc getXmlRpcFileTransferStatuses getFileTransferStatusesFromXmlRpc getXmlRpcProductPage getProductPageFromXmlRpc getXmlRpcComplexQuery getComplexQueryFromXmlRpc getXmlRpcQueryFilter getQueryFilterFromXmlRpc getXmlRpcFilterAlgor getFilterAlgorFromXmlRpc getXmlRpcQueryResults getQueryResultsFromXmlRpc getXmlRpcQueryResult getQueryResultFromXmlRpc getXmlRpcProduct getProductFromXmlRpc getProductListFromXmlRpc getXmlRpcProductList getXmlRpcProductTypeList getProductTypeListFromXmlRpc getXmlRpcProductType getProductTypeFromXmlRpc getXmlRpcTypeExtractors getXmlRpcExtractorSpec getXmlRpcTypeHandlers getXmlRpcTypeHandler getTypeExtractorsFromXmlRpc getExtractorSpecFromXmlRpc getTypeHandlersFromXmlRpc getTypeHandlerFromXmlRpc getPropertiesFromXmlRpc getXmlRpcProperties getXmlRpcReferences getReferencesFromXmlRpc getXmlRpcReference getReferenceFromXmlRpc getReferenceFromXmlRpcHashtable getXmlRpcElementListHashtable getXmlRpcElementList getElementListFromXmlRpc getXmlRpcElement getXmlRpcElementHashTable getElementFromXmlRpc getXmlRpcQuery getQueryFromXmlRpc getXmlRpcQueryCriteriaList getQueryCriteriaListFromXmlRpc getXmlRpcQueryCriteria getQueryCriteriaFromXmlRpcexamples.MajorMinorVersionTypeHandler getCatalogValue getOrigValuesystem.FileManager LOG logger catalog repositoryManager dataTransfer transferStatusTracker expandProductMet configurationManager setCatalog refreshConfigAndPolicy transferringProduct getCurrentFileTransfer getCurrentFileTransfers getProductPctTransferred getRefPctTransferred removeProductTransferStatus isTransferComplete pagedQuery getFirstPage getLastPage getNextPage getPrevPage addProductType setProductTransferStatus getNumProducts getTopNProducts getTopNProductsByProductType hasProduct getProductTypes getProductReferences getProductById getProductByName getProductsByProductType getElementsByProductType getElementById getElementByName complexQuery getProductTypeByName getProductTypeById updateMetadata ingestProduct retrieveFile transferFile moveProduct removeFile modifyProduct removeProduct catalogProduct addMetadata runExtractors addProductReferences setProductType query getReducedMetadata getMetadata getOrigValues getCatalogValues getCatalogQuery applyFilterToResults sortQueryResultList buildProductMetadata loadConfiguration setDataTransferaction.GetPrevPageCliAction productTypeName currentPageNum execute setProductTypeName setCurrentPageNumextractors.AbstractFilemgrMetExtractor LOG configuration extractMetadata configure doExtract doConfigure validateProduct addMetadataIfUndefined merge getProductFile getRootRefPathstructs.FreeTextQueryCriteria serialVersionUID elementName values noiseWords noiseWordHash getValues setValue addValue addFreeText getElementName setElementName toStringingest.CmdLineIngester LOG main readProdFilesFromStdinsystem.AvroFileManagerServer logger port server fileManager startUp loadConfiguration shutdown isAlive setCatalog refreshConfigAndPolicy transferringProduct getCurrentFileTransfer getCurrentFileTransfers getProductPctTransferred getRefPctTransferred removeProductTransferStatus isTransferComplete pagedQuery getFirstPage getLastPage getNextPage getPrevPage setProductTransferStatus getNumProducts getTopNProductsByProductType getTopNProducts hasProduct getMetadata getReducedMetadata getProductTypes getProductReferences getProductById getProductByName getProductsByProductType getElementsByProductType getElementById getElementByName complexQuery query getProductTypeByName getProductTypeById updateMetadata addProductType catalogProduct addMetadata addProductReferences ingestProduct retrieveFile transferFile moveProduct removeFile modifyProduct removeProduct getCatalogValues getOrigValues getCatalogQuerysystem.FileManagerClientMain mainmetadata.CoreMetKeysrepository.XMLRepositoryManager productTypeHomeUris productTypeMap LOG addProductType modifyProductType removeProductType getProductTypeById getProductTypeByName getProductTypes saveProductTypes loadProductTypes getDocumentRootaction.SqlQueryCliAction query getQuery setQuerydatatransfer.DataTransfersolr.UUIDProductIdGenerator generateIdtools.MetadataBasedProductMover DOUBLE pathSpec fmgrClient LOG moveProducts locationsMatch moveProducts mainaction.GetProductPercentTransferredCliAction productId productTypeName execute setProductId setProductTypeNametools.CASAnalyzer stopSet factory STOP_WORDS reader createComponents tokenStreamsvalidation.ScienceDataValidationLayerFactory dataSource createValidationLayeraction.AbstractQueryCliAction sortBy outputFormat delimiter filterAlgor startDateTimeMetKey endDateTimeMetKey priorityMetKey versionConverter execute getQuery setSortBy setOutputFormat setDelimiter setFilterAlgor setStartDateTimeMetKey setEndDateTimeMetKey setPriorityMetKey setVersionConverterrepository.ScienceDataRepositoryManagerFactory dataSource createRepositoryManagersolr.ProductSerializeringest.StdIngester LOG fmClient clientTransferServiceFactory ingest ingest ingest hasProduct check checkOrSetFileManager setFileManager getProductType hasProduct closesolr.CompleteProduct product metadata getProduct setProduct getMetadata setMetadatacatalog.LenientDataSourceCatalog LOG isoFormat dbFormat addMetadata getMetadataTypes removeMetadata getMetadata populateProductMetadata getReducedMetadata addMetadataValue removeMetadataValue getResultListSize getSqlQuerytools.QueryTool freeTextBlock client LOG query parseQuery generateCASQuery main safeGetProductTypes performSqlQuery exittools.MetadataDumper LOG fmClient FILENAME PRODUCT_NAME getMetadata writeMetFileToDir dumpMetadata dumpMetadata mainvalidation.XMLValidationLayerFactory dirList LOG createValidationLayerversioning.ProductTypeMetVersioner createDataStoreReferencesexceptions.VersioningException serialVersionUIDvalidation.ValidationLayerFactoryingest.RemoteableCachesolr.DefaultProductSerializer LOG getMimeType serialize serialize deserialize serialize addKeyValueToMap generateInsertDocuments generateUpdateDocuments toDoc encodeIndexField encodeUpdateField deserialize parseXml deserializeSingleValueField deserializeMultiValueFieldvalidation.ScienceDataValidationLayer LOG ds addElement addElementToProductType getElementById getElementByName getElements getElements modifyElement removeElement removeElementFromProductTypedatatransfer.RemoteDataTransferFactory chunkSize LOG createDataTransfersolr.SolrCatalogFactory solrUrl productIdGenerator productSerializer LOG configure createCatalogversioning.DateTimeVersioner LOG createDataStoreReferences addProductDateTimeToReferencessystem.CommonsXmlRpcTransport url client userAgentHeader http11 gzip rgzip creds method timeout connecttimeout password user auth sendXmlRpc setHttp11 setGzip setRGzip setUserAgent setTimeout setConnectionTimeout setBasicAuthentication setBasicAuthentication endClientRequestrpc.XmlRpcFileManagerServerFactory port getPort setPort createFileManagerServertype.TypeHandler elementName setElementName getElementName preQueryHandle handleQueryCriteria postGetMetadataHandle preAddMetadataHandle handleRangeQueryCriteria handleTermQueryCriteriautil.QueryUtils getQueryResultsAsString getQueryResultsAsStringtools.ExpImpCatalog sourceClient destClient srcCatalog destCatalog ensureUnique LOG doExpImport doExpImport exportTypeToDest exportProductsToDest isEnsureUnique setEnsureUnique main typesExist typeInList safeHasProductTypeByNameexceptions.CatalogException serialVersionUIDstructs.BooleanQueryCriteria AND OR NOT serialVersionUID operator terms addTerm getTerms setOperator getOperator getElementName setElementName toStringtools.OptimizeLuceneCatalog DOUBLE INT reader config catalogPath mergeFactor LOG doOptimize mainsystem.XmlRpcFileManagerClient client LOG fileManagerUrl dataTransfer refreshConfigAndPolicy isAlive transferringProduct removeProductTransferStatus isTransferComplete moveProduct modifyProduct removeProduct getCurrentFileTransfer getCurrentFileTransfers getProductPctTransferred getRefPctTransferred pagedQuery getFirstPage getLastPage getNextPage getPrevPage addProductType hasProduct getNumProducts getTopNProducts getTopNProducts setProductTransferStatus addProductReferences addMetadata updateMetadata catalogProduct getMetadata getReducedMetadata removeFile retrieveFile transferFile getProductsByProductType getElementsByProductType getElementById getElementByName getElementByName complexQuery query getProductTypeByName getProductTypeById getProductTypes getProductReferences getProductById getProductByName ingestProduct getCatalogValues getOrigValues getCatalogQuery main getFileManagerUrl setFileManagerUrl getDataTransfer setDataTransfer closesolr.SolrCatalog productSerializer productIdGenerator solrClient LOG addMetadata removeMetadata addProduct modifyProduct removeProduct setProductTransferStatus addProductReferences getProductById getProductByName getProductReferences getProducts getProducts getProducts getProductsByProductType getMetadata getReducedMetadata query getTopNProducts pagedQuery getFirstPage getLastProductPage getNextPage getPrevPage getProductsFromDocument getTopNProducts getValidationLayer getNumProducts getCompleteProductById getCompleteProductByName extractCompleteProduct newProductPagedatatransfer.TransferStatusTracker currentProductTransfers catalog LOG getCurrentFileTransfer transferringProduct getCurrentFileTransfers getPctTransferred getPctTransferred removeProductTransferStatus isTransferComplete getBytesTransferred quietGetReferences isDirdatatransfer.InPlaceDataTransferFactory createDataTransfertools.SolrIndexer LOG SOLR_INDEXER_CONFIG SOLR_URL FILEMGR_URL ACCESS_KEY ACCESS_URL PRODUCT_NAME config server fmUrl solrUrl solrFormat delete commit optimize getSolrDocument indexMetFile indexProductTypes safeFirstPage indexAll indexProduct indexProductByName indexProduct deleteProduct deleteProductByName deleteProductFromIndex performSubstitution formatDate buildCommandLine main readProductIdsFromStdinexamples.DateTimeExpandMetExtractor FULL_DATE_TIME_KEY FULL_DATE_TIME_FORMAT YEAR_KEY MONTH_KEY DAY_KEY HOUR_KEY MINUTE_KEY SECOND_KEY fullDateTimeKey fullDateTimeFormat yearKey monthKey dayKey hourKey minuteKey secondKey doConfigure doExtract getKey getFullDateTime createDateFieldaction.AddProductTypeCliAction productTypeId productTypeName productTypeDescription fileRepositoryPath versioner execute setProductTypeId setProductTypeName setProductTypeDescription setFileRepositoryPath setVersionerexceptions.ConnectionException serialVersionUIDtools.ProductTypeDocTool xslFilePath outputDirPath doProductTypeDoc mainaction.GetProductByNameCliAction productName getProduct setProductNamesystem.FileManagerServerMain mainvalidation.DataSourceValidationLayer LOG dataSource quoteFields addElement modifyElement removeElement addElementToProductType removeElementFromProductType addParentToProductType removeParentFromProductType getElements getElements getElementById getElementByNameversioning.ConfigurableMetadataBasedFileVersioner BASE_PROPERTY ALL properties createDataStoreReferences getFilePathSpecingest.LocalCacheFactory createCacherpc.AvroFileManagerServerFactory port createFileManagerServer getPort setPortaction.GetCurrentTransfersCliAction executecatalog.Catalogaction.FileManagerCliAction fmc getUrl getClient setClientstructs.FileTransferStatus fileRef bytesTransferred parentProduct getBytesTransferred setBytesTransferred getFileRef setFileRef getParentProduct setParentProduct computePctTransferredutil.AvroTypeFactory logger getAvroReference getReference getAvroExtractorSpec getExtractorSpec getAvroMetadata getMetadata getAvroTypeHandler getTypeHandler getAvroProductType getProductType getAvroProduct getProduct getAvroFileTransferStatus getFileTransferStatus getAvroQueryCriteria getQueryCriteria getAvroQuery getQuery getAvroProductPage getProductPage getAvroElement getElement getAvroQueryResult getQueryResult getAvroFilterAlgor getFilterAlgor getAvroQueryFilter getQueryFilter getAvroComplexQuery getComplexQuery getReferences getAvroReferencesversioning.BasicVersioner LOG createDataStoreReferencescatalog.MappedDataSourceCatalog typeMap addMetadata addProductReferences getMetadata getProductReferences modifyProduct removeMetadata removeProduct pagedQuery query getProductTypeTableNameextractors.FilemgrMetExtractorutil.GenericFileManagerObjectFactory LOG getDataTransferServiceFromFactory getRepositoryManagerServiceFromFactory getCatalogServiceFromFactory getValidationLayerFromFactory getCacheFromFactory getVersionerFromClassName getExtractorFromClassName getTypeHandlerFromClassName getFilterAlgorFromClassName getVersionConverterFromClassNamesystem.FileManagerClientdatatransfer.LocalDataTransferer LOG client setFileManagerUrl transferProduct retrieveProduct deleteProduct main copyDirToDir moveDirToProductRepo moveFilesToProductRepo copyFilesToDir moveFile copyFile quietNotifyTransferProduct quietNotifyProductTransferComplete finalizeaction.DeleteProductByIdCliAction productId getProductToDelete setProductIdexceptions.FileManagerExceptionaction.AbstractDeleteProductCliAction execute getProductToDeleterepository.DataSourceRepositoryManager dataSource LOG addProductType modifyProductType removeProductType getProductTypeById getProductTypeByName getProductTypesingest.RmiCacheFactory rmiCacheServerUrn createCacheversioning.Versioneraction.AbstractGetProductCliAction execute getProducttype.ValueReplaceTypeHandler postGetMetadataHandle preAddMetadataHandle handleRangeQueryCriteria handleTermQueryCriteria getCatalogValue getOrigValuedatatransfer.DataTransferFactorystructs.Query criteria getCriteria setCriteria addCriterion toStringaction.DeleteProductByNameCliAction productName getProductToDelete setProductNameaction.HasProductCliAction productName execute setProductNamerepository.XMLRepositoryManagerFactory productTypeDirList LOG createRepositoryManagerexceptions.DataTransferException serialVersionUIDauth.Dispatchersolr.Parameters ID NS PRODUCT_ID PRODUCT_NAME PRODUCT_STRUCTURE PRODUCT_TRANSFER_STATUS PRODUCT_RECEIVED_TIME PRODUCT_TYPE_NAME PRODUCT_TYPE_ID REFERENCE_ORIGINAL REFERENCE_DATASTORE REFERENCE_FILESIZE REFERENCE_MIMETYPE ROOT ROOT_REFERENCE_ORIGINAL ROOT_REFERENCE_DATASTORE ROOT_REFERENCE_FILESIZE ROOT_REFERENCE_MIMETYPE SOLR_DATE_FORMAT SOLR_DATE_TIME_FORMATTER MIME_TYPE_XML MIME_TYPE_JSON PAGE_SIZE NULLaction.GetNextPageCliAction productTypeName currentPageNum execute setProductTypeName setCurrentPageNumvalidation.DataSourceValidationLayerFactory dataSource quoteFields createValidationLayerdatatransfer.LocalDataTransferFactory createDataTransferaction.DumpMetadataCliAction productId outputDir execute setProductId setOutputDir generateFilenamestructs.ExtractorSpec className configuration getClassName setClassName getConfiguration setConfigurationutil.RpcCommunicationFactory LOG getClientFactoryName setPror createClient createClient createServerversioning.MetadataBasedFileVersioner filePathSpec LOG flatProducts createDataStoreReferences parseFilePathSpec getFilePathSpec setFilePathSpec isFlatProducts setFlatProductssolr.NameProductIdGenerator generateIdutil.DbStructFactory getProductType getProduct getProduct getReference getElement getParent toScienceDataElement toScienceDataProduct toScienceDataProductTypecatalog.ScienceDataCatalog dataSource validationLayer pageSize LOG addMetadata createGranule createParameter commitQuery addProduct addProductReferences getMetadata getNumProducts getProductById getProductByName getProductReferences getProducts getProductsByProductType getReducedMetadata getTopNProducts getTopNProducts getValidationLayer modifyProduct query removeMetadata removeProduct setProductTransferStatus getFirstPage getLastProductPage getNextPage getPrevPage pagedQuery paginateQuery getResultListSize getStartDateTime getEndDateTime getMetadataSubsetexamples.TikaAutoDetectExtractor doConfigure doExtract getMetadataFromTika transformdatatransfer.S3DataTransfererFactory BUCKET_NAME_PROPERTY REGION_PROPERTY ACCESS_KEY_PROPERTY SECRET_KEY_PROPERTY ENCRYPT_PROPERTY createDataTransferaction.GetFilePercentTransferredCliAction origRef execute setOrigRef getUrirepository.DataSourceRepositoryManagerFactory dataSource createRepositoryManagervalidation.ValidationLayersystem.AvroFileManagerClient logger client proxy fileManagerUrl dataTransfer refreshConfigAndPolicy isAlive transferringProduct removeProductTransferStatus isTransferComplete moveProduct modifyProduct removeProduct getCurrentFileTransfer getCurrentFileTransfers getProductPctTransferred getRefPctTransferred pagedQuery getFirstPage getLastPage getNextPage getPrevPage addProductType hasProduct getNumProducts getTopNProducts getTopNProducts setProductTransferStatus addProductReferences addMetadata updateMetadata catalogProduct getMetadata getReducedMetadata removeFile retrieveFile transferFile getProductsByProductType getElementsByProductType getElementById getElementByName complexQuery query getProductTypeByName getProductTypeById getProductTypes getProductReferences getProductById getProductByName ingestProduct getCatalogValues getOrigValues getCatalogQuery getFileManagerUrl setFileManagerUrl getDataTransfer setDataTransfer closecatalog.DataSourceCatalog INT dataSource LOG validationLayer fieldIdStringFlag pageSize productIdString orderedValues PRODUCT_CACHE cacheUpdateMinutes addMetadata removeMetadata addProduct modifyProduct removeProduct setProductTransferStatus addProductReferences getProductById getProductByName getProductReferences getProducts getProductsByProductType getMetadata getReducedMetadata query getTopNProducts getTopNProducts getValidationLayer addMetadataValue removeMetadataValue getNumProducts getFirstPage getLastProductPage getNextPage getPrevPage pagedQuery getResultListSize stillFresh getProductsFromCache flagCacheUpdate getProductsByProductTypeCached paginateQuery getSqlQuery updateReferences quoteItrepository.RepositoryManagerFactorycatalog.CatalogFactoryaction.IngestProductCliAction LOG productName productStructure productTypeName metadataFile dataTransferer references execute setProductName setProductStructure setProductTypeName setMetadataFile setDataTransferer setReferences getUrisolr.ProductIdGeneratorstructs.RangeQueryCriteria serialVersionUID elementName startValue endValue inclusive getStartValue setStartValue getEndValue setEndValue getInclusive setInclusive getElementName setElementName toStringstructs.TermQueryCriteria serialVersionUID elementName value getValue setValue getElementName setElementName toStringexamples.FileAttributesExtractor BASIC_FILE_ATTRIBUTES POSIX_FILE_ATTRIBUTES attributes LOG doExtract doConfigure getMetadataFromFileAttributesaction.GetLastPageCliAction productTypeName execute setProductTypeNameaction.LuceneQueryCliAction FREE_TEXT_BLOCK query reducedProductTypes reducedMetadataKeys getQuery setQuery setReducedProductTypes setReducedMetadataKeys parseQuery generateCASQueryquery.ComplexQuery reducedProductTypeNames reducedMetadata queryFilter sortByMetKey toStringResultFormat getReducedProductTypeNames setReducedProductTypeNames getReducedMetadata setReducedMetadata getQueryFilter setQueryFilter getSortByMetKey setSortByMetKey getToStringResultFormat setToStringResultFormattools.DumpDbElementsToXml mainaction.GetFirstPageCliAction productTypeName execute setProductTypeNameaction.RetrieveFilesCliAction productId productName dt destination execute setProductId setProductName setDataTransferFactory setDestinationconv.VersionConverterversioning.InPlaceVersioner LOG createDataStoreReferencesingest.Ingesteringest.CachedIngester cache LOG hasProduct hasProduct resynsc init closeversioning.SingleFileBasicVersioner FILENAME_FIELD LOG createDataStoreReferencesingest.RmiCache LOG rmiCacheServer clear contains setFileManager size sync sync setUniqueElementProductTypeNames sync setUniqueElementName getFileManagerUrlingest.RmiCacheServer LOG serialVersionUID cache uniqueElementName uniqueElementProductTypeNames reg launchServer launchServer stopServer clear contains setFileManager size sync sync sync getFileManagerUrl getUniqueElementProductTypeNames setUniqueElementProductTypeNames getUniqueElementName setUniqueElementName syncWith launchRmiServercatalog.ScienceDataCatalogFactory dataSource pageSize createCatalogexceptions.ValidationLayerException serialVersionUIDstructs.ProductType productTypeId name description productRepositoryPath versioner typeMetadata extractors handlers getDescription setDescription getName setName getProductTypeId setProductTypeId getProductRepositoryPath setProductRepositoryPath getVersioner setVersioner getTypeMetadata setTypeMetadata getExtractors setExtractors getHandlers setHandlers toString blankProductTypevalidation.XMLValidationLayer LOG productTypeElementMap subToSuperMap elementMap xmlFileDirUris addElement modifyElement removeElement addElementToProductType removeElementFromProductType getElements getElements getElementById getElementByName getElements getSubToSuperMap addParentForProductType removeParentForProductType saveElementsAndMappings loadElements loadProductTypeMap getDocumentRootstructs.Reference LOG origReference dataStoreReference fileSize mimeType mimeTypeRepository STREAM_REFERENCE_DELIMITER getDataStoreReference setDataStoreReference getOrigReference setOrigReference getFileSize setFileSize getMimeType setMimeType setMimeType toStringstructs.Element elementId elementName dcElement description hashCode getDCElement setDCElement getElementId setElementId getElementName setElementName getDescription setDescription blankElementextractors.CoreMetExtractor namespaceAware elementNs nsSeparator nsReplaceElements doExtract doConfigure isNsReplaceexamples.FilenameRegexMetExtractor filenamePattern metadataKeys doConfigure doExtractfilter.TimeEvent startTime priority setPriority getPriority getStartTime getEndTime getDuration equals toString happenAtSameTime compareTo hashCodecatalog.MappedDataSourceCatalogFactory typeMap TYPE_MAP_KEY createCatalogfilter.ObjectTimeEvent timeObj getTimeObject equals toString hashCodequery.QueryResultComparator sortByMetKey getSortByMetKey setSortByMetKey compareexceptions.QueryFormulationException serialVersionUIDconv.AsciiSortableVersionConverter convertToPriorityauth.SecureWebServer dispatchers execute addDispatcherrepository.ScienceDataRepositoryManager LOG dataSource addProductType getProductTypeById getProductTypeByName getProductTypes modifyProductType removeProductTypecatalog.LuceneCatalogFactory VAL VAL1 VAL2 VAL3 indexFilePath config validationLayer pageSize writeLockTimeOut commitLockTimeOut mergeFactor lenientFields LOG createCatalogingest.CacheFactoryrepository.RepositoryManagerexceptions.RepositoryManagerException serialVersionUIDsystem.FileManagerServerversioning.VersioningUtils LOG DIR_FILTER FILE_FILTER getReferencesFromDir getURIsFromDir createBasicDataStoreRefsHierarchical createBasicDataStoreRefsFlat createBasicDataStoreRefsStream createDataStoreRefStream addRefsFromUris getAbsolutePathFromUri quietGetFileSizeFromUriversioning.DirectoryProductVersioner createDataStoreReferencesquery.QueryFilter startDateTimeMetKey filterAlgor converter getStartDateTimeMetKey setStartDateTimeMetKey getEndDateTimeMetKey setEndDateTimeMetKey getPriorityMetKey setPriorityMetKey getFilterAlgor setFilterAlgor getConverter setConverteraction.GetNumProductsCliAction productTypeName execute setProductTypeNameaction.GetProductTypeByNameCliAction productTypeName execute setProductTypeNamefilter.FilterAlgor epsilon setEpsilon getEpsilon filterEventssolr.SolrClient solrUrl LOG index delete queryProductById queryProductByName queryProductsByDate queryProductsByDateAndType commit query doGet doPost doHttp buildUpdateUrl buildSelectUrlutil.Paginationrpc.XmlRpcFileManagerClientFactory url testConnection setUrl getUrl setTestConnection getTestConnection createFileManagerClientsystem.XmlRpcFileManager webServerPort catalog repositoryManager dataTransfer LOG webServer transferStatusTracker expandProductMet configurationManager configurationListener setCatalog isAlive refreshConfigAndPolicy transferringProduct transferringProductCore getCurrentFileTransferCore getCurrentFileTransfers getProductPctTransferred getProductPctTransferredCore getRefPctTransferred getRefPctTransferredCore removeProductTransferStatus removeProductTransferStatusCore isTransferComplete isTransferCompleteCore pagedQuery pagedQueryCore getFirstPage getFirstPageCore getLastPage getLastPageCore getNextPage getNextPageCore getPrevPage getPrevPageCore addProductType addProductTypeCore setProductTransferStatus setProductTransferStatusCore getNumProducts getNumProductsCore getTopNProducts getTopNProducts getTopNProductsCore hasProduct getMetadata getMetadataCore getReducedMetadata getReducedMetadataCore getProductTypes getProductReferences getProductReferencesCore getProductById getProductByName getProductsByProductType getProductsByProductTypeCore getElementsByProductType getElementById getElementByName complexQuery complexQueryCore query queryCore getProductTypeByName getProductTypeById updateMetadata updateMetadataCore catalogProduct catalogProductCore addMetadata addMetadataCore addProductReferencesCore addProductReferences ingestProduct ingestProductCore retrieveFile transferFile moveProduct moveProductCore removeFile modifyProduct modifyProductCore removeProduct removeProduct getCatalogValues getCatalogValuesCore getOrigValues getOrigValuesCore getCatalogQuery getCatalogQueryCore main shutdown catalogProduct addMetadata runExtractors addProductReferences setProductType query getReducedMetadata getMetadata getOrigValues getCatalogValues getCatalogQuery applyFilterToResults sortQueryResultList buildProductMetadata loadConfigurationtools.RangeQueryTester startFieldName endFieldName startFieldStartValue startFieldEndValue endFieldStartValue endFieldEndValue indexPath LOG reader doRangeQuery main getEndFieldName setEndFieldName getIndexPath setIndexPath getStartFieldName setStartFieldName getEndFieldEndValue setEndFieldEndValue getEndFieldStartValue setEndFieldStartValue getStartFieldEndValue setStartFieldEndValue getStartFieldStartValue setStartFieldStartValuestructs.ProductPage pageNum totalPages pageSize pageProducts numOfHits getPageNum setPageNum getPageProducts setPageProducts getPageSize setPageSize getTotalPages setTotalPages isLastPage isFirstPage getNumOfHits setNumOfHits blankPage toStringauth.Result value ARGS getValueversioning.AcquisitionDateVersioner filePathSpec ACQUISITION_DATE ACQ_DATE_FORMAT START_DATE_TIME createDataStoreReferencestools.DeleteProduct LOG client commit remove isCommit setCommit main readProdIdsFromStdin finalizeutil.SqlParser LOG parseSqlQueryMethod parseSqlQuery parseSqlWhereClause unparseSqlQuery getInfixCriteriaString getInfixCriteriaString stripOutSqlDefinition getSqlStatementArgs createFilter toPostFix parseStatement listToString mainrpc.FileManagerClientFactorymetadata.FileAttributesMetKeysrpc.AvroFileManagerClientFactory url testConnection setUrl getUrl setTestConnection getTestConnection createFileManagerClientsolr.QueryResponse numFound start completeProducts setNumFound setStart setResults getNumFound getStart getCompleteProducts getProductsstructs.QueryCriteria getElementName setElementName toStringingest.AbstractCacheServerFactory rangeQueryElementName rangeStartDateTime rangeEndDateTime uniqueElementName productTypeNames fmUrl createCachetools.ProductDumper LOG fmClient FILENAME PRODUCT_NAME getProduct writeProductFileToDir getMetadata dumpProduct dumpProduct mainingest.Cacheingest.LocalCache uniqueElements uniqueElementName uniqueElementProductTypeNames rangeQueryElementName startOfQuery endOfQuery fm LOG sync sync clear size contains setFileManager sync getFileManagerUrl liveHasProduct getUniqueElementName setUniqueElementName getUniqueElementProductTypeNames setUniqueElementProductTypeNames getProductsOverDateRange getProducts getValueForMetadata finalizeaction.GetCurrentTransferCliAction executetools.CatalogSearch LOG client freeTextBlock productFilter PostQuery PostQuery setFilter removeFilter ListProducts listElements listElements printHelp ParseQuery GenerateCASQuery CommandParser mainingest.RmiCacheServerFactory createCache createRemoteCache mainmetadata.ProductMetKeyssystem.XmlRpcFileManagerServer port webServer fileManager isAlive setCatalog startUp loadConfiguration refreshConfigAndPolicy transferringProduct getCurrentFileTransfer getCurrentFileTransfers getRefPctTransferred removeProductTransferStatus isTransferComplete pagedQuery getFirstPage getLastPage getNextPage getProductPctTransferred getPrevPage addProductType setProductTransferStatus getNumProducts getTopNProducts getTopNProducts hasProduct getMetadata getReducedMetadata getProductTypes getProductReferences getProductById getProductByName getProductsByProductType getElementsByProductType getElementById getElementByName complexQuery query getProductTypeByName getProductTypeById updateMetadata catalogProduct addMetadata addProductReferences ingestProduct retrieveFile transferFile moveProduct removeFile modifyProduct removeProduct getCatalogValues getOrigValues getCatalogQuery shutdown \ No newline at end of file +"Reference", +"Versioner", +"Element", +"Metadata", +"ProductType", +"Product", +"exceptions CacheException", +"examples MimeTypeExtractor", +"structs Product", +"exceptions IngestException", +"datatransfer S3DataTransferer", +"action GetProductByIdCliAction", +"query QueryResult", +"catalog LuceneCatalog", +"datatransfer RemoteDataTransferer", +"catalog DataSourceCatalogFactory", +"rpc FileManagerServerFactory", +"util XmlStructFactory", +"datatransfer InPlaceDataTransferer", +"examples FinalFileLocationExtractor", +"util XmlRpcStructFactory", +"examples MajorMinorVersionTypeHandler", +"system FileManager", +"action GetPrevPageCliAction", +"extractors AbstractFilemgrMetExtractor", +"structs FreeTextQueryCriteria", +"ingest CmdLineIngester", +"system AvroFileManagerServer", +"system FileManagerClientMain", +"metadata CoreMetKeys", +"repository XMLRepositoryManager", +"action SqlQueryCliAction", +"datatransfer DataTransfer", +"solr UUIDProductIdGenerator", +"tools MetadataBasedProductMover", +"action GetProductPercentTransferredCliAction", +"tools CASAnalyzer", +"validation ScienceDataValidationLayerFactory", +"action AbstractQueryCliAction", +"repository ScienceDataRepositoryManagerFactory", +"solr ProductSerializer", +"ingest StdIngester", +"solr CompleteProduct", +"catalog LenientDataSourceCatalog", +"tools QueryTool", +"tools MetadataDumper", +"validation XMLValidationLayerFactory", +"versioning ProductTypeMetVersioner", +"exceptions VersioningException", +"validation ValidationLayerFactory", +"ingest RemoteableCache", +"solr DefaultProductSerializer", +"validation ScienceDataValidationLayer", +"datatransfer RemoteDataTransferFactory", +"solr SolrCatalogFactory", +"versioning DateTimeVersioner", +"system CommonsXmlRpcTransport", +"rpc XmlRpcFileManagerServerFactory", +"type TypeHandler", +"util QueryUtils", +"tools ExpImpCatalog", +"exceptions CatalogException", +"structs BooleanQueryCriteria", +"tools OptimizeLuceneCatalog", +"system XmlRpcFileManagerClient", +"solr SolrCatalog", +"datatransfer TransferStatusTracker", +"datatransfer InPlaceDataTransferFactory", +"tools SolrIndexer", +"examples DateTimeExpandMetExtractor", +"action AddProductTypeCliAction", +"exceptions ConnectionException", +"tools ProductTypeDocTool", +"action GetProductByNameCliAction", +"system FileManagerServerMain", +"validation DataSourceValidationLayer", +"versioning ConfigurableMetadataBasedFileVersioner", +"ingest LocalCacheFactory", +"rpc AvroFileManagerServerFactory", +"action GetCurrentTransfersCliAction", +"catalog Catalog", +"action FileManagerCliAction", +"structs FileTransferStatus", +"util AvroTypeFactory", +"versioning BasicVersioner", +"catalog MappedDataSourceCatalog", +"extractors FilemgrMetExtractor", +"util GenericFileManagerObjectFactory", +"system FileManagerClient", +"datatransfer LocalDataTransferer", +"action DeleteProductByIdCliAction", +"exceptions FileManagerException", +"action AbstractDeleteProductCliAction", +"repository DataSourceRepositoryManager", +"ingest RmiCacheFactory", +"versioning Versioner", +"action AbstractGetProductCliAction", +"type ValueReplaceTypeHandler", +"datatransfer DataTransferFactory", +"structs Query", +"action DeleteProductByNameCliAction", +"action HasProductCliAction", +"repository XMLRepositoryManagerFactory", +"exceptions DataTransferException", +"auth Dispatcher", +"solr Parameters", +"action GetNextPageCliAction", +"validation DataSourceValidationLayerFactory", +"datatransfer LocalDataTransferFactory", +"action DumpMetadataCliAction", +"structs ExtractorSpec", +"util RpcCommunicationFactory", +"versioning MetadataBasedFileVersioner", +"solr NameProductIdGenerator", +"util DbStructFactory", +"catalog ScienceDataCatalog", +"examples TikaAutoDetectExtractor", +"datatransfer S3DataTransfererFactory", +"action GetFilePercentTransferredCliAction", +"repository DataSourceRepositoryManagerFactory", +"validation ValidationLayer", +"system AvroFileManagerClient", +"catalog DataSourceCatalog", +"repository RepositoryManagerFactory", +"catalog CatalogFactory", +"action IngestProductCliAction", +"solr ProductIdGenerator", +"structs RangeQueryCriteria", +"structs TermQueryCriteria", +"examples FileAttributesExtractor", +"action GetLastPageCliAction", +"action LuceneQueryCliAction", +"query ComplexQuery", +"tools DumpDbElementsToXml", +"action GetFirstPageCliAction", +"action RetrieveFilesCliAction", +"conv VersionConverter", +"versioning InPlaceVersioner", +"ingest Ingester", +"ingest CachedIngester", +"versioning SingleFileBasicVersioner", +"ingest RmiCache", +"ingest RmiCacheServer", +"catalog ScienceDataCatalogFactory", +"exceptions ValidationLayerException", +"structs ProductType", +"validation XMLValidationLayer", +"structs Reference", +"structs Element", +"extractors CoreMetExtractor", +"examples FilenameRegexMetExtractor", +"filter TimeEvent", +"catalog MappedDataSourceCatalogFactory", +"filter ObjectTimeEvent", +"query QueryResultComparator", +"exceptions QueryFormulationException", +"conv AsciiSortableVersionConverter", +"auth SecureWebServer", +"repository ScienceDataRepositoryManager", +"catalog LuceneCatalogFactory", +"ingest CacheFactory", +"repository RepositoryManager", +"exceptions RepositoryManagerException", +"system FileManagerServer", +"versioning VersioningUtils", +"versioning DirectoryProductVersioner", +"query QueryFilter", +"action GetNumProductsCliAction", +"action GetProductTypeByNameCliAction", +"filter FilterAlgor", +"solr SolrClient", +"util Pagination", +"rpc XmlRpcFileManagerClientFactory", +"system XmlRpcFileManager", +"tools RangeQueryTester", +"structs ProductPage", +"auth Result", +"versioning AcquisitionDateVersioner", +"tools DeleteProduct", +"util SqlParser", +"rpc FileManagerClientFactory", +"metadata FileAttributesMetKeys", +"rpc AvroFileManagerClientFactory", +"solr QueryResponse", +"structs QueryCriteria", +"ingest AbstractCacheServerFactory", +"tools ProductDumper", +"ingest Cache", +"ingest LocalCache", +"action GetCurrentTransferCliAction", +"tools CatalogSearch", +"ingest RmiCacheServerFactory", +"metadata ProductMetKeys", +"system XmlRpcFileManagerServer" \ No newline at end of file diff --git a/src/main/resources/cdtocode/Apache OODT File Manager-VSMLSI.xls b/src/main/resources/cdtocode/Apache OODT File Manager-VSMLSI.xls new file mode 100644 index 0000000000000000000000000000000000000000..3008abc03df4df747c76943d3e79fa7a951ff44f Binary files /dev/null and b/src/main/resources/cdtocode/Apache OODT File Manager-VSMLSI.xls differ diff --git a/src/main/resources/cdtocode/Apache OODT File Manager-VSMLSIdata.xls b/src/main/resources/cdtocode/Apache OODT File Manager-VSMLSIdata.xls index ebf4bd8c389d691a8fd282149f67cdcf94ac436e..c3eebd35431494a4d90685baa0c9697c15648c06 100644 Binary files a/src/main/resources/cdtocode/Apache OODT File Manager-VSMLSIdata.xls and b/src/main/resources/cdtocode/Apache OODT File Manager-VSMLSIdata.xls differ diff --git a/src/main/resources/cdtocode/Apache OODT File Manager-noadd.xls b/src/main/resources/cdtocode/Apache OODT File Manager-noadd.xls new file mode 100644 index 0000000000000000000000000000000000000000..e383063dc2cc91c4bb25ed0fd77e7dcf6412e483 Binary files /dev/null and b/src/main/resources/cdtocode/Apache OODT File Manager-noadd.xls differ diff --git a/src/main/resources/cdtocode/Apache OODT File Manager-nodoc.xls b/src/main/resources/cdtocode/Apache OODT File Manager-nodoc.xls new file mode 100644 index 0000000000000000000000000000000000000000..9f6dee7a53b3368d7a36f2684939f223650eb82d Binary files /dev/null and b/src/main/resources/cdtocode/Apache OODT File Manager-nodoc.xls differ diff --git a/src/main/resources/cdtocode/Apache OODT File Manager-nolog.xls b/src/main/resources/cdtocode/Apache OODT File Manager-nolog.xls new file mode 100644 index 0000000000000000000000000000000000000000..1145b621adb63de6d99db1caf4702bab1e39e9b9 Binary files /dev/null and b/src/main/resources/cdtocode/Apache OODT File Manager-nolog.xls differ diff --git a/src/main/resources/cdtocode/Apache OODT File Manager-vsmlsiraw.xls b/src/main/resources/cdtocode/Apache OODT File Manager-vsmlsiraw.xls new file mode 100644 index 0000000000000000000000000000000000000000..ad175d6a7a32d4d2d5334d30823e02dc7c39955b Binary files /dev/null and b/src/main/resources/cdtocode/Apache OODT File Manager-vsmlsiraw.xls differ diff --git a/src/main/resources/cdtocode/Hadoop HDFS-VSMLSI.txt b/src/main/resources/cdtocode/Hadoop HDFS-VSMLSI.txt new file mode 100644 index 0000000000000000000000000000000000000000..eb83bc487f872eb2061ac32220701d7d61659198 --- /dev/null +++ b/src/main/resources/cdtocode/Hadoop HDFS-VSMLSI.txt @@ -0,0 +1,1234 @@ +"NameNode", +"DataNode", +"Tools", +"Balancer", +"Protocol", +"Security", +"Client", +"Common", +"blockmanagement BlockPlacementPolicy", +"client package-info", +"store StateStoreCacheUpdateService", +"util CyclicIteration", +"util ReferenceCountMap", +"datanode BlockPoolSliceStorage", +"erasurecode StripedBlockWriter", +"protocol RemoveMountTableEntryResponse", +"fs WebHdfs", +"records StateStoreVersion", +"protocol LayoutFlags", +"ha AbstractNNFailoverProxyProvider", +"fs SWebHdfs", +"aliasmap InMemoryAliasMapProtocol", +"datanode ReplicaWaitingToBeRecovered", +"namenode CheckpointSignature", +"router ErasureCoding", +"web SWebHdfsFileSystem", +"blockmanagement CacheReplicationMonitor", +"blockmanagement BlockPlacementPolicyRackFaultTolerant", +"fsdataset FsVolumeReference", +"tools DebugAdmin", +"router RouterRpcMonitor", +"datamodel DiskBalancerCluster", +"protocol NNHAStatusHeartbeat", +"protocol DataNodeUsageReportUtil", +"client HdfsDataInputStream", +"impl BlockReaderLocalLegacy", +"protocol SlowPeerReports", +"server JournalMetrics", +"blockmanagement BlockReconstructionWork", +"namenode FSTreeTraverser", +"util BestEffortLongFile", +"tools HDFSConcat", +"inotify EventBatchList", +"namenode FSDirSymlinkOp", +"ha NameNodeHAProxyFactory", +"sps BlockMovementAttemptFinished", +"impl FsDatasetUtil", +"protocol AddMountTableEntryRequest", +"startupprogress Phase", +"nfs3 DFSClientCache", +"protocolPB ReconfigurationProtocolServerSideUtils", +"offlineImageViewer FileDistributionVisitor", +"impl FsVolumeImpl", +"namenode ErasureCodingPolicyManager", +"datanode StorageLocation", +"hdfs SWebHdfsDtFetcher", +"fsdataset DataNodeVolumeMetrics", +"aliasmap InMemoryLevelDBAliasMapServer", +"namenode NameNodeHttpServer", +"datatransfer Sender", +"datanode BlockPoolManager", +"router ConnectionContext", +"oauth2 OAuth2ConnectionConfigurator", +"protocolPB JournalProtocolServerSideTranslatorPB", +"server CheckUploadContentTypeFilter", +"server JournalNodeMXBean", +"util package-info", +"protocol JournalInfo", +"datanode DataXceiverServer", +"blockmanagement BlockPlacementPolicyWithUpgradeDomain", +"ha ActiveState", +"resources ConcatSourcesParam", +"pb EnterSafeModeRequestPBImpl", +"hdfs ExceptionLastSeen", +"offlineEditsViewer OfflineEditsBinaryLoader", +"hdfs DFSStripedOutputStream", +"protocol RouterHeartbeatResponse", +"protocol OpenFileEntry", +"protocol CorruptFileBlocks", +"protocolPB PBHelperClient", +"protocol StorageReport", +"router NamenodeHeartbeatService", +"offlineImageViewer PBImageDelimitedTextWriter", +"protocol GetSafeModeResponse", +"planner Planner", +"namenode FSImagePreTransactionalStorageInspector", +"web URLConnectionFactory", +"resolver MultipleDestinationMountTableResolver", +"protocol SnapshotException", +"namenode BackupJournalManager", +"ha BootstrapStandby", +"protocol NamenodeHeartbeatRequest", +"sps BlocksMovementsStatusHandler", +"namenode Content", +"offlineImageViewer FSImageLoader", +"store package-info", +"blockmanagement HostFileManager", +"pb RouterHeartbeatRequestPBImpl", +"pb GetDisabledNameservicesResponsePBImpl", +"service FileSystemAccessException", +"server GetJournalEditServlet", +"namenode AclStorage", +"pb EnableNameserviceResponsePBImpl", +"mover Mover", +"pb DisableNameserviceResponsePBImpl", +"offlineEditsViewer XmlEditsVisitor", +"web ByteRangeInputStream", +"router RouterClient", +"datanode BPServiceActor", +"namenode EncryptionFaultInjector", +"datanode CorruptMetaHeaderException", +"namenode BackupNode", +"router RouterMetrics", +"protocol SnapshotDiffReportListing", +"blockmanagement OutOfLegacyGenerationStampsException", +"util ByteArrayManager", +"datanode BlockReceiver", +"tools DelegationTokenFetcher", +"web SSLConnectionConfigurator", +"protocolPB InterQJournalProtocolServerSideTranslatorPB", +"blockmanagement DatanodeDescriptor", +"resources StringParam", +"ha ObserverReadProxyProviderWithIPFailover", +"federation RouterAdmin", +"hdfs ClientContext", +"checker VolumeCheckResult", +"blockmanagement DatanodeAdminManager", +"protocol ExtendedBlock", +"namenode Namesystem", +"protocolPB DatanodeProtocolPB", +"namenode FSPermissionChecker", +"servlet FileSystemReleaseFilter", +"erasurecode StripedReconstructionInfo", +"namenode FSDirMkdirOp", +"impl DisabledNameserviceStoreImpl", +"snapshot SnapshotDiff", +"protocol UpdateMountTableEntryRequest", +"sps ItemInfo", +"pb StateStoreVersionPBImpl", +"wsrs InputStreamEntity", +"protocol DatanodeLifelineProtocol", +"namenode LogsPurgeable", +"namenode NNStorage", +"hdfs StripeReader", +"records DisabledNameservice", +"offlineImageViewer OfflineImageViewer", +"protocol HdfsFileStatus", +"datanode ReportBadBlockAction", +"fsdataset RoundRobinVolumeChoosingPolicy", +"datanode BlockScanner", +"resolver FederationNamenodeServiceState", +"router IsRouterActiveServlet", +"server JournalNodeSyncer", +"blockmanagement DatanodeStats", +"protocolPB NamenodeProtocolTranslatorPB", +"wsrs JSONMapProvider", +"protocol DatanodeRegistration", +"hdfs BlockMissingException", +"resources OffsetParam", +"blockmanagement BlockManagerFaultInjector", +"checker AbstractFuture", +"metrics StateStoreMetrics", +"startupprogress StartupProgress", +"protocol ECTopologyVerifierResult", +"protocolPB DatanodeLifelineProtocolServerSideTranslatorPB", +"oauth2 CredentialBasedAccessTokenProvider", +"util CombinedHostsFileReader", +"impl LevelDBFileRegionAliasMap", +"order OrderedResolver", +"blockmanagement SequentialBlockGroupIdGenerator", +"pb MountTablePBImpl", +"metrics OutlierDetector", +"offlineImageViewer IgnoreSnapshotException", +"protocol NamenodeRegistration", +"util Canceler", +"namenode CheckpointFaultInjector", +"snapshot SnapshotStatsMXBean", +"checker Checkable", +"namenode FSDirDeleteOp", +"namenode NameNode", +"protocol VolumeFailureSummary", +"metrics FederationRPCPerformanceMonitor", +"protocol GetNamespaceInfoResponse", +"resolver RemoteLocation", +"planner NodePlan", +"protocolPB JournalProtocolTranslatorPB", +"resources ModificationTimeParam", +"service Scheduler", +"namenode FSImageStorageInspector", +"mount RpcProgramMountd", +"tools ECAdmin", +"impl FsDatasetCache", +"servlet HostnameFilter", +"protocol CheckpointCommand", +"resources ExcludeDatanodesParam", +"client QuorumJournalManager", +"impl StateStoreFileSystemImpl", +"blockmanagement Host2NodesMap", +"snapshot DirectoryWithSnapshotFeature", +"hadoop FileSystemAccessService", +"blockmanagement BlocksMap", +"datanode DiskBalancerWorkStatus", +"connectors ClusterConnector", +"datanode DiskBalancerWorkItem", +"server JournalNode", +"shortcircuit DfsClientShm", +"router PeriodicService", +"checker DatasetVolumeChecker", +"protocol StorageReceivedDeletedBlocks", +"protocol DatanodeStorage", +"offlineEditsViewer OfflineEditsXmlLoader", +"namenode INodeDirectory", +"store DisabledNameserviceStore", +"web URLDispatcher", +"impl MemoryMappableBlockLoader", +"pb UpdateMountTableEntryRequestPBImpl", +"impl BlockReaderFactory", +"namenode ContentSummaryComputationContext", +"namenode FSDirWriteFileOp", +"ha HAProxyFactory", +"resources DelegationParam", +"lang XException", +"namenode FSEditLogLoader", +"hdfs ClientGSIContext", +"protocol DatanodeID", +"protocol RollingUpgradeStatus", +"hdfs DeprecatedUTF8", +"protocol package-info", +"protocol GetDisabledNameservicesRequest", +"wsrs ParametersProvider", +"resources NamenodeAddressParam", +"pb GetRouterRegistrationsRequestPBImpl", +"protocol InterDatanodeProtocol", +"util Diff", +"namenode VersionInfoMXBean", +"protocolPB AliasMapProtocolPB", +"datatransfer IOStreamPair", +"router RouterQuotaUpdateService", +"common HttpGetFailedException", +"protocolPB JournalProtocolPB", +"hdfs ReplicaAccessorBuilder", +"impl AddBlockPoolException", +"hdfs DFSInputStream", +"resources AclPermissionParam", +"command ExecuteCommand", +"ha RemoteNameNodeInfo", +"util LongBitFormat", +"namenode RedundantEditLogInputStream", +"pb RouterHeartbeatResponsePBImpl", +"store StateStoreService", +"balancer Balancer", +"protocol SlowDiskReports", +"startupprogress package-info", +"planner PlannerFactory", +"protocol HdfsPathHandle", +"datanode DNConf", +"protocol RemoteEditLog", +"hdfs DFSUtilClient", +"common StorageInfo", +"namenode NotReplicatedYetException", +"namenode EditLogInputException", +"namenode XAttrPermissionFilter", +"resources CreateFlagParam", +"blockmanagement PendingDataNodeMessages", +"protocol EnableNameserviceResponse", +"records MembershipStats", +"pb NamenodeHeartbeatRequestPBImpl", +"datanode ReplicaAlreadyExistsException", +"util ByteArray", +"records MembershipState", +"pb GetNamespaceInfoResponsePBImpl", +"server HttpFSReleaseFilter", +"protocol ErasureCodingPolicyState", +"protocol RequestInfo", +"namenode XAttrFeature", +"offlineImageViewer PBImageTextWriter", +"hdfs DFSOutputStream", +"sasl DataEncryptionKeyFactory", +"protocolPB ClientDatanodeProtocolPB", +"sps StoragePolicySatisfier", +"util ByteBufferOutputStream", +"namenode NameNodeMXBean", +"protocol DatanodeInfo", +"resources DestinationParam", +"planner GreedyPlanner", +"datanode Replica", +"namenode NameNodeResourcePolicy", +"hdfs ReaderStrategy", +"client SegmentRecoveryComparator", +"common HdfsServerConstants", +"protocol BlockReportContext", +"client QuorumOutputStream", +"client HdfsAdmin", +"protocolPB InterDatanodeProtocolPB", +"server ServiceException", +"balancer KeyManager", +"fsdataset ReplicaOutputStreams", +"namenode StartupProgressServlet", +"block DataEncryptionKey", +"pb GetNamespaceInfoRequestPBImpl", +"datanode ReplicaUnderRecovery", +"impl StateStoreZooKeeperImpl", +"pb MembershipStatsPBImpl", +"metrics FederationRPCMBean", +"client DfsPathCapabilities", +"common ECTopologyVerifier", +"blockmanagement DatanodeStatistics", +"util StripedBlockUtil", +"hdfs ExtendedBlockId", +"nfs3 OpenFileCtxCache", +"blockmanagement BlockUnderConstructionFeature", +"hdfs KeyProviderCache", +"impl PmemMappedBlock", +"protocol NodeRegistration", +"top TopAuditLogger", +"driver StateStoreSerializer", +"protocol GetRouterRegistrationsResponse", +"client BlockReportOptions", +"blockmanagement BlockIdManager", +"namenode StoragePolicySummary", +"impl StateStoreBaseImpl", +"namenode CheckpointConf", +"protocol CachePoolEntry", +"checker AsyncChecker", +"hdfs DFSClient", +"protocol FsPermissionExtension", +"metrics DataNodeMetrics", +"protocolPB NamenodeProtocolServerSideTranslatorPB", +"connectors package-info", +"namenode EditLogBackupInputStream", +"block BlockPoolTokenSecretManager", +"store StateStoreUnavailableException", +"router RouterNamenodeProtocol", +"resources XAttrNameParam", +"protocolPB ClientDatanodeProtocolServerSideTranslatorPB", +"namenode FSImageCompression", +"hdfs StripedDataStreamer", +"client IPCLoggerChannel", +"impl CacheStats", +"namenode AclEntryStatusFormat", +"util Check", +"blockmanagement ExcessRedundancyMap", +"namenode Checkpointer", +"namenode INodeFile", +"resolver FederationNamespaceInfo", +"resources OverwriteParam", +"resources PutOpParam", +"datanode DataNodeFaultInjector", +"nfs3 AsyncDataService", +"pb LeaveSafeModeRequestPBImpl", +"resources UriFsPathParam", +"impl StateStoreSerializerPBImpl", +"inotify MissingEventsException", +"metrics ReplicatedBlocksMBean", +"server Server", +"resources NewLengthParam", +"pb RouterStatePBImpl", +"impl FsDatasetImpl", +"datanode ReplicaInfo", +"driver StateStoreRecordOperations", +"datanode FileIoProvider", +"snapshot AbstractINodeDiffList", +"records Query", +"namenode INodeAttributeProvider", +"protocol LeaveSafeModeResponse", +"namenode BackupImage", +"shortcircuit DfsClientShmManager", +"datanode BPOfferService", +"protocolPB ReconfigurationProtocolTranslatorPB", +"impl StateStoreFileBaseImpl", +"hdfs HDFSPolicyProvider", +"namenode RetryStartFileException", +"impl MountTableStoreImpl", +"wsrs StringParam", +"protocol BlockIdCommand", +"datatransfer PacketHeader", +"block ExportedBlockKeys", +"namenode NameNodeRpcServer", +"namenode NameNodeFormatException", +"shortcircuit ClientMmap", +"datanode ReplicaBeingWritten", +"blockmanagement HeartbeatManager", +"sps SPSService", +"protocol QuotaByStorageTypeExceededException", +"blockmanagement BlockStatsMXBean", +"protocol ZoneReencryptionStatus", +"command HelpCommand", +"namenode SecondaryNameNodeInfoMXBean", +"resources HttpOpParam", +"protocol NamenodeCommand", +"balancer MovedBlocks", +"resolver MountTableResolver", +"snapshot FSImageFormatPBSnapshot", +"resources BooleanParam", +"server FSOperations", +"datanode ReplicaHandler", +"client HdfsDataOutputStream", +"namenode AuditLogger", +"protocol BlocksStorageMoveAttemptFinished", +"net Peer", +"wsrs ExceptionProvider", +"protocol ReencryptionStatusIterator", +"datatransfer WhitelistBasedTrustedChannelResolver", +"protocol EncryptionZone", +"common FileRegion", +"snapshot DiffListByArrayList", +"resources PostOpParam", +"common JspHelper", +"pb EnableNameserviceRequestPBImpl", +"blockmanagement BlockPlacementStatusDefault", +"blockmanagement CombinedHostFileManager", +"resources RenewerParam", +"protocolPB ClientDatanodeProtocolTranslatorPB", +"namenode CachePool", +"sasl InvalidMagicNumberException", +"shortcircuit ShortCircuitReplicaInfo", +"offlineImageViewer ImageVisitor", +"hdfs DFSPacket", +"ha IPFailoverProxyProvider", +"pb RemoveMountTableEntryRequestPBImpl", +"protocol GetNamespaceInfoRequest", +"impl FsVolumeList", +"protocol RollingUpgradeInfo", +"protocol CacheDirective", +"common BlockAlias", +"instrumentation InstrumentationService", +"protocol NamenodeHeartbeatResponse", +"ha HAState", +"protocolPB QJournalProtocolServerSideTranslatorPB", +"sps BlockDispatcher", +"protocol DatanodeVolumeInfo", +"hdfs CannotObtainBlockLengthException", +"hdfs DataStreamer", +"blockmanagement HostSet", +"datatransfer package-info", +"blockmanagement BlockPlacementPolicyDefault", +"blockmanagement BlockPlacementStatus", +"util DataTransferThrottler", +"web SimpleHttpProxyHandler", +"namenode SaveNamespaceContext", +"startupprogress StepTracking", +"protocol DSQuotaExceededException", +"ha EditLogTailer", +"shortcircuit ShortCircuitShm", +"resources LengthParam", +"namenode FSEditLogAsync", +"namenode FsckServlet", +"hdfs DFSStripedInputStream", +"datanode ReplicaNotFoundException", +"pb GetSafeModeRequestPBImpl", +"datatransfer BlockConstructionStage", +"resources XAttrValueParam", +"hdfs NameNodeProxies", +"offlineEditsViewer OfflineEditsVisitor", +"util ECPolicyLoader", +"protocolPB InterDatanodeProtocolTranslatorPB", +"impl ReplicaCachingGetSpaceUsed", +"namenode FSDirSatisfyStoragePolicyOp", +"namenode FSImageFormat", +"server HttpFSServerWebServer", +"net PeerServer", +"namenode FSImageTransactionalStorageInspector", +"blockmanagement BlockReportLeaseManager", +"namenode ReencryptionUpdater", +"pb MembershipStatePBImpl", +"datanode BlockMetadataHeader", +"impl MemoryMappedBlock", +"impl MembershipStoreImpl", +"blockmanagement CorruptReplicasMap", +"block InvalidBlockTokenException", +"namenode IsNameNodeActiveServlet", +"delegation DelegationTokenSecretManager", +"tools GetGroups", +"protocol HdfsNamedFileStatus", +"protocolPB InterDatanodeProtocolServerSideTranslatorPB", +"fs Hdfs", +"blockaliasmap package-info", +"hdfs ReadStatistics", +"offlineEditsViewer BinaryEditsVisitor", +"protocol GetRouterRegistrationsRequest", +"impl MappableBlockLoader", +"pb GetRouterRegistrationRequestPBImpl", +"pb RemoveMountTableEntryResponsePBImpl", +"protocol QJournalProtocol", +"command Command", +"namenode FSDirSnapshotOp", +"wsrs IntegerParam", +"offlineImageViewer PBImageXmlWriter", +"impl SnapshotDiffReportGenerator", +"nfs3 RpcProgramNfs3", +"order AvailableSpaceResolver", +"protocol HdfsConstants", +"protocol FenceResponse", +"client AsyncLoggerSet", +"datamodel DiskBalancerVolumeSet", +"impl VolumeFailureInfo", +"inotify Event", +"namenode CacheManager", +"hdfs DFSUtil", +"datanode ChunkChecksum", +"blockmanagement DatanodeManager", +"pb UpdateNamenodeRegistrationRequestPBImpl", +"protocol RouterHeartbeatRequest", +"snapshot SnapshotFSImageFormat", +"router ConnectionPoolId", +"metrics FSDatasetMBean", +"util CombinedHostsFileWriter", +"window RollingWindowManager", +"protocol GetSafeModeRequest", +"store StateStoreConnectionMonitorService", +"namenode HdfsAuditLogger", +"diskbalancer DiskBalancerException", +"protocol EnterSafeModeRequest", +"webhdfs ParameterParser", +"webhdfs ExceptionHandler", +"impl LeaseRenewer", +"nfs3 Nfs3HttpServer", +"impl ReplicaMap", +"sps package-info", +"fsdataset FsVolumeSpi", +"protocolPB ReconfigurationProtocolUtils", +"hdfs HAUtil", +"metrics StateStoreMBean", +"protocol StripedBlockInfo", +"ha RequestHedgingProxyProvider", +"util IOUtilsClient", +"connectors ConnectorFactory", +"nfs3 Nfs3", +"router ConnectionManager", +"impl NativePmemMappableBlockLoader", +"blockmanagement BlockInfoContiguous", +"datanode BPServiceActorAction", +"namenode INodesInPath", +"offlineImageViewer XmlImageVisitor", +"server JournalNodeHttpServer", +"snapshot FileWithSnapshotFeature", +"namenode FSNDNCacheOp", +"namenode INodeDirectoryAttributes", +"blockmanagement LocatedBlockBuilder", +"blockmanagement PendingRecoveryBlocks", +"protocol BlockRecoveryCommand", +"namenode FSDirAttrOp", +"resources NameSpaceQuotaParam", +"wsrs JSONProvider", +"namenode INodeMap", +"datanode FaultInjectorFileIoEvents", +"protocol GetDisabledNameservicesResponse", +"protocolPB RouterAdminProtocolTranslatorPB", +"datanode ErrorReportAction", +"erasurecode StripedReader", +"net TcpPeerServer", +"datamodel DiskBalancerDataNode", +"ha ConfiguredFailoverProxyProvider", +"datanode DataNodeMXBean", +"resources AccessTimeParam", +"namenode EditLogFileInputStream", +"protocol LeaveSafeModeRequest", +"protocol ReconfigurationProtocol", +"records QueryResult", +"net DFSNetworkTopology", +"impl FsDatasetFactory", +"blockmanagement LowRedundancyBlocks", +"nfs3 Nfs3Metrics", +"protocol JournalProtocol", +"blockmanagement ReplicaUnderConstruction", +"protocolPB DatanodeLifelineProtocolClientSideTranslatorPB", +"datanode DataNode", +"metrics DataNodePeerMetrics", +"datatransfer DataTransferProtoUtil", +"shortcircuit DomainSocketFactory", +"namenode NNUpgradeUtil", +"driver StateStoreDriver", +"datatransfer DataTransferProtocol", +"protocol SnapshotDiffReport", +"router RouterMetricsService", +"shortcircuit ShortCircuitCache", +"nfs3 PrivilegedNfsGatewayStarter", +"order package-info", +"router RouterHttpServer", +"tools StoragePolicyAdmin", +"security GroupsService", +"offlineImageViewer OfflineImageReconstructor", +"server Service", +"datamodel package-info", +"snapshot Snapshot", +"hdfs ReplicaAccessor", +"checker StorageLocationChecker", +"command QueryCommand", +"protocol ErasureCodingPolicyInfo", +"tools DFSAdmin", +"protocol UnresolvedPathException", +"blockmanagement BlockStoragePolicySuite", +"impl BlockPoolSlice", +"protocol UnregisteredNodeException", +"pb GetRouterRegistrationResponsePBImpl", +"balancer Dispatcher", +"pb AddMountTableEntryResponsePBImpl", +"datatransfer PipelineAck", +"hdfs DFSInotifyEventInputStream", +"datanode LocalReplica", +"metrics FederationMBean", +"metrics NamenodeBeanMetrics", +"impl PmemVolumeManager", +"resources RenameOptionSetParam", +"protocolPB ReconfigurationProtocolServerSideTranslatorPB", +"blockmanagement HostConfigManager", +"protocol LastBlockWithStatus", +"pb AddMountTableEntryRequestPBImpl", +"protocol JournalOutOfSyncException", +"offlineImageViewer DelimitedImageVisitor", +"resources NoRedirectParam", +"blockmanagement NumberReplicas", +"protocol DatanodeAdminProperties", +"offlineImageViewer OfflineImageViewerPB", +"datanode ShortCircuitRegistry", +"protocol EncryptionZoneIterator", +"planner package-info", +"protocol ServerCommand", +"oauth2 OAuth2Constants", +"blockmanagement SequentialBlockIdGenerator", +"namenode EditLogBackupOutputStream", +"offlineImageViewer FSImageHandler", +"blockmanagement BlockManagerSafeMode", +"ha StandbyState", +"namenode NameNodeUtils", +"metrics TopMetrics", +"datanode DataXceiver", +"namenode FSNamesystem", +"datanode ReplicaInPipeline", +"namenode GlobalStateIdContext", +"utils package-info", +"blockmanagement ReplicationWork", +"order LocalResolver", +"protocol SystemErasureCodingPolicies", +"web RestCsrfPreventionFilterHandler", +"offlineEditsViewer TeeOutputStream", +"blockmanagement UnresolvedTopologyException", +"driver package-info", +"server JournalNodeRpcServer", +"datanode DirectoryScanner", +"namenode AclTransformation", +"wsrs LongParam", +"namenode NameNodeLayoutVersion", +"resources StartAfterParam", +"offlineImageViewer ImageLoader", +"resolver NamenodeStatusReport", +"protocol CacheDirectiveStats", +"order RandomResolver", +"blockmanagement BlockCollection", +"datanode DataNodeLayoutVersion", +"impl NativePmemMappedBlock", +"protocol BalancerProtocols", +"client QuorumException", +"datanode VolumeScannerCBInjector", +"tools DFSHAAdmin", +"namenode CachedBlock", +"snapshot SnapshotDiffListingInfo", +"namenode XAttrStorage", +"webhdfs WebHdfsHandler", +"impl StateStoreFileImpl", +"pb DisabledNameservicePBImpl", +"namenode ReencryptionHandler", +"protocol BlockStorageMovementCommand", +"protocolPB RouterAdminProtocolServerSideTranslatorPB", +"sps BlockStorageMovementNeeded", +"blockmanagement DatanodeStorageInfo", +"impl TextFileRegionAliasMap", +"pb package-info", +"order HashFirstResolver", +"store StateStoreUtils", +"blockmanagement BlockToMarkCorrupt", +"resources EnumParam", +"net DomainPeerServer", +"datanode DiskFileCorruptException", +"pb PBRecord", +"blockmanagement SlowPeerTracker", +"protocol CachePoolInfo", +"resolver FileSubclusterResolver", +"snapshot SnapshotDiffInfo", +"connectors JsonNodeConnector", +"sps StoragePolicySatisfyManager", +"impl ReservedSpaceCalculator", +"offlineEditsViewer OfflineEditsVisitorFactory", +"util FoldedTreeSet", +"protocol RecoveryInProgressException", +"metrics ECBlockGroupsMBean", +"protocol LayoutVersion", +"blockmanagement BlockManager", +"namenode LeaseExpiredException", +"sps Context", +"impl RamDiskAsyncLazyPersistService", +"resources PermissionParam", +"mount Mountd", +"service FileSystemAccess", +"snapshot LsSnapshottableDir", +"wsrs ShortParam", +"resolver package-info", +"hdfs WebHdfsDtFetcher", +"protocol ClientDatanodeProtocol", +"protocol LocatedBlocks", +"util EnumCounters", +"fsdataset FsDatasetSpi", +"startupprogress StartupProgressView", +"util XMLUtils", +"service Instrumentation", +"datanode SecureDataNodeStarter", +"namenode FSEditLogOpCodes", +"fs HdfsBlockLocation", +"nfs3 OpenFileCtx", +"impl BlockReaderLocal", +"ha ObserverReadProxyProvider", +"order HashResolver", +"checker package-info", +"oauth2 AccessTokenProvider", +"protocolPB package-info", +"namenode BackupState", +"block BlockKey", +"store RecordStore", +"resources IntegerParam", +"protocol GetNamenodeRegistrationsRequest", +"util LightWeightLinkedSet", +"util MD5FileUtils", +"servlet MDCFilter", +"impl RamDiskReplicaLruTracker", +"protocol NamespaceInfo", +"pb GetNamenodeRegistrationsResponsePBImpl", +"pb LeaveSafeModeResponsePBImpl", +"protocol ECBlockGroupStats", +"protocol ProvidedStorageLocation", +"resolver ActiveNamenodeResolver", +"protocol BlockListAsLongs", +"protocolPB AliasMapProtocolServerSideTranslatorPB", +"protocol GetRouterRegistrationRequest", +"metrics BlockReaderIoProvider", +"client HdfsClientConfigKeys", +"nfs3 Nfs3Utils", +"util RwLock", +"connectors DBNameNodeConnector", +"common IncorrectVersionException", +"ha WrappedFailoverProxyProvider", +"protocol RemoteEditLogManifest", +"planner MoveStep", +"router RouterClientProtocol", +"router RouterWebHdfsMethods", +"protocol FSLimitException", +"protocol ClientProtocol", +"hdfs DistributedFileSystem", +"ha ReadOnly", +"blockmanagement SlowDiskTracker", +"protocol CacheDirectiveInfo", +"protocol JournalNotFormattedException", +"protocol GetRouterRegistrationResponse", +"diskbalancer DiskBalancerConstants", +"resources Param", +"datanode DiskBalancer", +"erasurecode package-info", +"webhdfs DataNodeUGIProvider", +"resources SnapshotNameParam", +"server BaseService", +"fsdataset ReplicaInputStreams", +"hdfs AddBlockFlag", +"balancer NameNodeConnector", +"lang RunnableCallable", +"nfs3 WriteManager", +"router RouterServiceState", +"tools JMXGet", +"hdfs RemotePeerFactory", +"net DFSTopologyNodeImpl", +"sps BlockMovementStatus", +"namenode FSDirAppendOp", +"util LightWeightHashSet", +"offlineImageViewer LsImageVisitor", +"protocol DatanodeLocalInfo", +"net DomainPeer", +"sasl SaslResponseWithNegotiatedCipherOption", +"wsrs Parameters", +"namenode INode", +"common Util", +"impl MappableBlockLoaderFactory", +"offlineEditsViewer OfflineEditsViewer", +"server HttpFSServerWebApp", +"impl package-info", +"util EnumDoubles", +"sps ExternalSPSContext", +"impl StateStoreSerializableImpl", +"protocol FinalizeCommand", +"resources DeleteOpParam", +"impl FsDatasetAsyncDiskService", +"command package-info", +"ha HAContext", +"namenode NameNodeStatusMXBean", +"web ParamFilter", +"snapshot DiffList", +"sps BlockMovementListener", +"protocolPB QJournalProtocolTranslatorPB", +"startupprogress Step", +"protocol AddErasureCodingPolicyResponse", +"router RouterStateManager", +"namenode UnsupportedActionException", +"balancer ExitStatus", +"namenode FSImageFormatProtobuf", +"namenode FSDirStatAndListingOp", +"client HttpFSUtils", +"datanode DatanodeUtil", +"protocol InterQJournalProtocol", +"shortcircuit ShortCircuitReplica", +"namenode InotifyFSEditLogOpTranslator", +"server HttpFSParametersProvider", +"startupprogress StartupProgressMetrics", +"protocol DisableNameserviceResponse", +"offlineImageViewer WebImageViewer", +"protocol CacheDirectiveIterator", +"datanode DataStorage", +"tools CacheAdmin", +"web JsonUtil", +"protocolPB DatanodeLifelineProtocolPB", +"wsrs EnumSetParam", +"pb UpdateMountTableEntryResponsePBImpl", +"snapshot FileDiffList", +"blockmanagement AvailableSpaceBlockPlacementPolicy", +"namenode INodeAttributes", +"common GenerationStamp", +"namenode DfsServlet", +"protocol UpdateMountTableEntryResponse", +"protocolPB ClientNamenodeProtocolPB", +"router RemoteMethod", +"resources TokenKindParam", +"common TokenVerifier", +"blockmanagement FSClusterStats", +"sps DatanodeCacheManager", +"resources NamenodeWebHdfsMethods", +"protocol DirectoryListing", +"pb GetMountTableEntriesResponsePBImpl", +"router RBFConfigKeys", +"store CachedRecordStore", +"tools GetConf", +"sps ExternalStoragePolicySatisfier", +"offlineImageViewer IndentedImageVisitor", +"resources TokenArgumentParam", +"impl RouterStoreImpl", +"resources XAttrSetFlagParam", +"protocol GetNamenodeRegistrationsResponse", +"hdfs UnknownCipherSuiteException", +"web JsonUtilClient", +"resources StorageTypeParam", +"web WebHdfsFileSystem", +"block BlockTokenIdentifier", +"offlineEditsViewer StatisticsEditsVisitor", +"protocol QuotaExceededException", +"protocol DataNodeUsageReport", +"oauth2 Utils", +"resolver MembershipNamenodeResolver", +"protocol SnapshotInfo", +"protocol NamenodeProtocols", +"web DatanodeHttpServer", +"sasl SaslParticipant", +"blockmanagement PendingReconstructionBlocks", +"top TopConf", +"erasurecode StripedBlockChecksumMd5CrcReconstructor", +"datatransfer InvalidEncryptionKeyException", +"nfs3 WriteCtx", +"protocolPB InterQJournalProtocolPB", +"sps BlockStorageMovementTracker", +"router FederationUtil", +"impl FsVolumeImplBuilder", +"impl CorruptFileBlockIterator", +"router ConnectionPool", +"protocol NamenodeProtocol", +"datanode IncrementalBlockReportManager", +"impl ExternalBlockReader", +"router Router", +"store MountTableStore", +"impl RamDiskReplicaTracker", +"service Groups", +"namenode INodeReference", +"datanode ReplicaBuilder", +"protocol NoECPolicySetException", +"protocol package-info", +"server JournaledEditsCache", +"namenode XAttrFormat", +"tools DFSZKFailoverController", +"blockmanagement StorageTypeStats", +"inotify EventBatch", +"tools DiskBalancerCLI", +"common MetricsLoggerTask", +"records RouterState", +"startupprogress Status", +"metrics DataNodeDiskMetrics", +"protocol EnterSafeModeResponse", +"resources EnumSetParam", +"tools DFSck", +"hdfs HdfsDtFetcher", +"sasl SaslDataTransferServer", +"checker ThrottledAsyncChecker", +"hdfs DFSMultipartUploaderFactory", +"datatransfer Op", +"hdfs PeerCache", +"erasurecode StripedBlockReconstructor", +"resources DoAsParam", +"router RouterQuotaUsage", +"namenode FSDirConcatOp", +"resources ExceptionHandler", +"store MembershipStore", +"pb FederationProtocolPBTranslator", +"protocol BlockType", +"protocol ReencryptionStatus", +"planner Step", +"protocolPB NamenodeProtocolPB", +"impl InMemoryLevelDBAliasMapClient", +"datanode LocalReplicaInPipeline", +"blockmanagement BlockPlacementPolicyWithNodeGroup", +"offlineImageViewer TextWriterImageVisitor", +"namenode NamenodeFsck", +"startupprogress PhaseTracking", +"namenode Quota", +"protocol BlockStoragePolicy", +"resources OldSnapshotNameParam", +"erasurecode StripedBlockChecksumReconstructor", +"router RemoteParam", +"wsrs Param", +"common StorageErrorReporter", +"namenode EncryptionZoneManager", +"datatransfer Receiver", +"fs CacheFlag", +"router SubClusterTimeoutException", +"impl PmemMappableBlockLoader", +"datanode UnexpectedReplicaStateException", +"webhdfs HdfsWriter", +"protocol HdfsLocatedFileStatus", +"namenode EditLogOutputStream", +"util Holder", +"client LoggerTooFarBehindException", +"order DestinationOrder", +"pb EnterSafeModeResponsePBImpl", +"tools CryptoAdmin", +"oauth2 AccessTokenTimer", +"impl BlockReaderUtil", +"namenode FSEditLog", +"datanode BlockSender", +"hdfs XAttrHelper", +"records BaseRecord", +"ha ClientHAProxyFactory", +"resources UserProvider", +"fsdataset AvailableSpaceVolumeChoosingPolicy", +"protocolPB ClientNamenodeProtocolServerSideTranslatorPB", +"conf NfsConfigKeys", +"protocol AlreadyBeingCreatedException", +"metrics package-info", +"util PersistentLongFile", +"common HttpPutFailedException", +"ha StandbyCheckpointer", +"namenode FileJournalManager", +"util ReadOnlyList", +"protocol BlockLocalPathInfo", +"util AtomicFileOutputStream", +"impl package-info", +"util ConstEnumCounters", +"hdfs HdfsConfiguration", +"offlineImageViewer FileDistributionCalculator", +"protocol DatanodeInfoWithStorage", +"resources UserParam", +"protocol BlocksWithLocations", +"protocol OpenFilesIterator", +"resolver NamenodePriorityComparator", +"protocol EnableNameserviceRequest", +"protocolPB InterQJournalProtocolTranslatorPB", +"server HttpFSServer", +"namenode SafeModeException", +"namenode IllegalReservedPathException", +"offlineEditsViewer OfflineEditsLoader", +"protocolPB DatanodeProtocolClientSideTranslatorPB", +"resources FsActionParam", +"namenode LeaseManager", +"scheduler SchedulerService", +"resolver MountTableManager", +"namenode NameNodeResourceChecker", +"namenode FileUnderConstructionFeature", +"sps ExternalSPSFilePathCollector", +"pb GetSafeModeResponsePBImpl", +"sasl DataTransferSaslUtil", +"router NameserviceManager", +"sps package-info", +"protocol RegisterCommand", +"metrics FSNamesystemMBean", +"router RouterRpcClient", +"blockmanagement BlockInfo", +"namenode EditLogFileOutputStream", +"protocol LocatedBlock", +"namenode FSDirRenameOp", +"blockmanagement ErasureCodingWork", +"datanode BlockRecoveryWorker", +"protocol HeartbeatResponse", +"hdfs BlockReader", +"hdfs NameNodeProxiesClient", +"namenode FSDirectory", +"namenode StreamLimiter", +"protocol DropSPSWorkCommand", +"namenode JournalSet", +"resources BufferSizeParam", +"datamodel DiskBalancerVolume", +"blockmanagement InvalidateBlocks", +"resolver PathLocation", +"protocol StorageBlockReport", +"protocol LocatedStripedBlock", +"fs XAttr", +"resources ECPolicyParam", +"router RouterHeartbeatService", +"sps FileCollector", +"resources GetOpParam", +"datatransfer TrustedChannelResolver", +"order RouterResolver", +"hdfs DFSOpsCountStatistics", +"command ReportCommand", +"blockmanagement BlockPlacementStatusWithNodeGroup", +"namenode FSImageFormatPBINode", +"metrics package-info", +"protocol DisableNameserviceRequest", +"impl package-info", +"block BlockTokenSelector", +"protocol ReplicaRecoveryInfo", +"client HttpsFSFileSystem", +"conf NfsConfiguration", +"blockmanagement BlockPlacementStatusWithUpgradeDomain", +"erasurecode StripedBlockReader", +"oauth2 ConfCredentialBasedAccessTokenProvider", +"namenode DefaultINodeAttributesProvider", +"namenode FSImageUtil", +"resources StorageSpaceQuotaParam", +"sps BlockStorageMovementAttemptedItems", +"delegation DelegationTokenIdentifier", +"snapshot SnapshotManager", +"pb DisableNameserviceRequestPBImpl", +"server JournalFaultInjector", +"startupprogress AbstractTracking", +"hdfs DFSClientFaultInjector", +"command CancelCommand", +"diskbalancer package-info", +"balancer BalancerParameters", +"erasurecode StripedReconstructor", +"namenode FSDirAclOp", +"resources RecursiveParam", +"client IPCLoggerChannelMetrics", +"namenode QuotaCounts", +"balancer Matcher", +"protocol UpdateNamenodeRegistrationRequest", +"protocolPB DatanodeProtocolServerSideTranslatorPB", +"blockaliasmap BlockAliasMap", +"startupprogress StepType", +"protocol CachePoolStats", +"pb GetDisabledNameservicesRequestPBImpl", +"hdfs HdfsKMSUtil", +"protocol DisallowedDatanodeException", +"erasurecode ErasureCodingWorker", +"namenode INodeWithAdditionalFields", +"protocol package-info", +"namenode NameCache", +"server Journal", +"snapshot AbstractINodeDiff", +"protocol AclException", +"wsrs BooleanParam", +"tools NNHAServiceTarget", +"web TokenAspect", +"datatransfer BlockPinningException", +"protocol DatanodeStorageReport", +"namenode SaveNamespaceCancelledException", +"offlineImageViewer DepthCounter", +"router Quota", +"pb NamenodeHeartbeatResponsePBImpl", +"namenode INodeSymlink", +"wsrs EnumParam", +"router DFSRouter", +"protocol UpdateNamenodeRegistrationResponse", +"blockmanagement BlockPlacementPolicies", +"datanode CachingStrategy", +"utils ConsistentHashRing", +"protocol RemoveMountTableEntryRequest", +"protocolPB ClientNamenodeProtocolTranslatorPB", +"resources LongParam", +"router RouterAdminServer", +"router RouterRpcServer", +"namenode INodeFileAttributes", +"metrics FederationMetrics", +"pb UpdateNamenodeRegistrationResponsePBImpl", +"resources ShortParam", +"snapshot FileDiff", +"sps BlockMoveTaskHandler", +"client AsyncLogger", +"protocol BlockECReconstructionCommand", +"server JNStorage", +"namenode EditLogInputStream", +"util ConfigurationUtils", +"protocol GetMountTableEntriesRequest", +"protocol ReceivedDeletedBlockInfo", +"datanode FinalizedProvidedReplica", +"protocol CachePoolIterator", +"resources OwnerParam", +"impl DfsClientConf", +"snapshot DirectorySnapshottableFeature", +"protocol AddMountTableEntryResponse", +"datanode VolumeScanner", +"namenode FSNamesystemLock", +"datanode ProfilingFileIoEvents", +"namenode FSImage", +"command PlanCommand", +"router package-info", +"namenode SerialNumberMap", +"namenode FSDirXAttrOp", +"metrics BlockReaderLocalMetrics", +"impl ProvidedVolumeImpl", +"namenode FSImageSerialization", +"sasl SaslDataTransferClient", +"server HttpFSAuthenticationFilter", +"block BlockTokenSecretManager", +"snapshot DirectoryDiffListFactory", +"federation package-info", +"protocol ErasureCodingPolicy", +"namenode FSDirTruncateOp", +"namenode EditsDoubleBuffer", +"wsrs ByteParam", +"blockmanagement ProvidedStorageMap", +"aliasmap InMemoryAliasMap", +"erasurecode StripedWriter", +"namenode AclFeature", +"protocol Block", +"tools AdminHelper", +"resources BlockSizeParam", +"hdfs FileChecksumHelper", +"namenode FSDirErasureCodingOp", +"router RemoteLocationContext", +"namenode FSEditLogOp", +"hdfs StatefulStripeReader", +"protocolPB QJournalProtocolPB", +"namenode QuotaByStorageTypeEntry", +"hdfs DFSConfigKeys", +"common InconsistentFSStateException", +"blockmanagement BlockInfoStriped", +"erasurecode StripedBlockChecksumCompositeCrcReconstructor", +"sps package-info", +"protocol BlockCommand", +"protocol DatanodeCommand", +"datanode BlockChecksumHelper", +"namenode SecondaryNameNode", +"resources TokenServiceParam", +"common Storage", +"datanode FSCachingGetSpaceUsed", +"web WebHdfsConstants", +"protocol CacheDirectiveEntry", +"pb package-info", +"impl MappableBlock", +"protocol BlockChecksumOptions", +"protocolPB RouterAdminProtocolPB", +"datanode ProvidedReplica", +"checker TimeoutFuture", +"client HdfsUtils", +"datatransfer ReplaceDatanodeOnFailure", +"hdfs UnknownCryptoProtocolVersionException", +"namenode DirectoryWithQuotaFeature", +"router RouterQuotaManager", +"store RouterStore", +"resources UnmaskedPermissionParam", +"namenode ImageServlet", +"net EncryptedPeer", +"offlineImageViewer ImageLoaderCurrent", +"client HttpFSFileSystem", +"resources ReplicationParam", +"metrics NameNodeMetrics", +"oauth2 ConfRefreshTokenBasedAccessTokenProvider", +"resolver FederationNamenodeContext", +"client QuorumCall", +"hdfs PositionStripeReader", +"namenode INodeId", +"router RouterPermissionChecker", +"servlet ServerWebApp", +"pb GetRouterRegistrationsResponsePBImpl", +"nfs3 OffsetRange", +"protocol ReplicatedBlockStats", +"hdfs DFSHedgedReadMetrics", +"namenode SafeMode", +"server HttpFSExceptionProvider", +"pb GetMountTableEntriesRequestPBImpl", +"protocol KeyUpdateCommand", +"namenode ContentCounts", +"pb GetNamenodeRegistrationsRequestPBImpl", +"namenode CheckableNameNodeResource", +"web AuthFilter", +"namenode NNStorageRetentionManager", +"datatransfer PacketReceiver", +"protocolPB PBHelper", +"store StateStoreCache", +"protocol BalancerBandwidthCommand", +"net NioInetPeer", +"namenode MetaRecoveryContext", +"namenode FSDirEncryptionZoneOp", +"snapshot DiffListBySkipList", +"ha InMemoryAliasMapFailoverProxyProvider", +"router RouterSafemodeService", +"fsdataset VolumeChoosingPolicy", +"namenode SerialNumberManager", +"protocol NSQuotaExceededException", +"protocolPB ReconfigurationProtocolPB", +"impl BlockReaderRemote", +"protocol DatanodeProtocol", +"namenode JournalManager", +"resources XAttrEncodingParam", +"namenode TransferFsImage", +"net BasicInetPeer", +"fsdataset LengthInputStream", +"protocol FencedException", +"protocol BlockChecksumType", +"delegation DelegationTokenSelector", +"protocol GetMountTableEntriesResponse", +"datatransfer BlackListBasedTrustedChannelResolver", +"web KerberosUgiAuthenticator", +"datanode FinalizedReplica", +"datanode BPServiceActorActionException", +"window RollingWindow", +"client CreateEncryptionZoneFlag", +"protocol SnapshottableDirectoryStatus", +"metrics FederationRPCMetrics", +"offlineImageViewer NameDistributionVisitor", +"sps ExternalSPSBlockMoveTaskHandler", +"protocolPB InMemoryAliasMapProtocolClientSideTranslatorPB", +"protocol RollingUpgradeException", +"records package-info", +"balancer BalancingPolicy", +"records MountTable", +"resources StoragePolicyParam", +"resources GroupParam", +"hdfs HAUtilClient", +"metrics DataNodeMetricHelper", +"oauth2 package-info", +"resources CreateParentParam", +"server ServerException", +"protocol SnapshotAccessControlException" \ No newline at end of file diff --git a/src/main/resources/cdtocode/Hadoop HDFS-VSMLSI.xls b/src/main/resources/cdtocode/Hadoop HDFS-VSMLSI.xls new file mode 100644 index 0000000000000000000000000000000000000000..9d3cefcfe7f54c5a2ddbe47e66e45b54087cb5b9 Binary files /dev/null and b/src/main/resources/cdtocode/Hadoop HDFS-VSMLSI.xls differ diff --git a/src/main/resources/cdtocode/Hadoop HDFS-VSMLSIdata.xls b/src/main/resources/cdtocode/Hadoop HDFS-VSMLSIdata.xls new file mode 100644 index 0000000000000000000000000000000000000000..7bee969e4d8e19dadcffcbf6ddf193020319c9ad Binary files /dev/null and b/src/main/resources/cdtocode/Hadoop HDFS-VSMLSIdata.xls differ diff --git a/src/main/resources/cdtocode/Hadoop HDFS-noadd.xls b/src/main/resources/cdtocode/Hadoop HDFS-noadd.xls new file mode 100644 index 0000000000000000000000000000000000000000..84f5ac46c01c1377b927fc7fd27c7e99c858a4e1 Binary files /dev/null and b/src/main/resources/cdtocode/Hadoop HDFS-noadd.xls differ diff --git a/src/main/resources/cdtocode/Hadoop HDFS-nodoc.xls b/src/main/resources/cdtocode/Hadoop HDFS-nodoc.xls new file mode 100644 index 0000000000000000000000000000000000000000..ac4afbfd7b12b74a2328adbb223a280d358fdfe1 Binary files /dev/null and b/src/main/resources/cdtocode/Hadoop HDFS-nodoc.xls differ diff --git a/src/main/resources/cdtocode/Hadoop HDFS-nolog.xls b/src/main/resources/cdtocode/Hadoop HDFS-nolog.xls new file mode 100644 index 0000000000000000000000000000000000000000..4d5c06bbf5cc616801c90dcf90caa7538d193eda Binary files /dev/null and b/src/main/resources/cdtocode/Hadoop HDFS-nolog.xls differ diff --git a/src/main/resources/cdtocode/Hadoop HDFS-vsmlsiraw.xls b/src/main/resources/cdtocode/Hadoop HDFS-vsmlsiraw.xls new file mode 100644 index 0000000000000000000000000000000000000000..009db3ac5c7ce2623a985f77ffe0ed9da16d6c23 Binary files /dev/null and b/src/main/resources/cdtocode/Hadoop HDFS-vsmlsiraw.xls differ diff --git a/src/main/resources/cdtocode/Hadoop MapReduce-VSMLSI.txt b/src/main/resources/cdtocode/Hadoop MapReduce-VSMLSI.txt new file mode 100644 index 0000000000000000000000000000000000000000..07e17fefc235037ffb72e71ba0aad33c6db001b9 --- /dev/null +++ b/src/main/resources/cdtocode/Hadoop MapReduce-VSMLSI.txt @@ -0,0 +1,966 @@ +"Partition", +"Mapper", +"OutputCommitter", +"Input", +"InputFormat", +"Output", +"OutputFormat", +"Reducer", +"Map", +"Reduce", +"InputSplit", +"Partitioner", +"authorize MRAMPolicyProvider", +"lib CombineFileSplit", +"rm RMHeartbeatHandler", +"mapred JvmContext", +"util CountersStrings", +"jobhistory HistoryEventHandler", +"jobhistory AMStartedEvent", +"mapreduce CounterGroup", +"mapred MapOutputCollector", +"input InvalidInputException", +"preemption NoopAMPreemptionPolicy", +"dao TaskAttemptsInfo", +"event TaskAttemptDiagnosticsUpdateEvent", +"db DBConfiguration", +"jobhistory HumanReadableHistoryViewerPrinter", +"pb GetDelegationTokenRequestPBImpl", +"output BindingPathOutputCommitter", +"commit CommitterEventType", +"mapred ProgressSplitsBlock", +"hs HSProxies", +"mapred MapTaskStatus", +"mapreduce JobSubmitter", +"mapred JobTracker", +"input SequenceFileInputFilter", +"chain ChainMapper", +"webapp HsView", +"util ConfigUtil", +"annotation Checkpointable", +"api HSClientProtocol", +"output LazyOutputFormat", +"jobhistory JobHistoryEventHandler", +"delegation DelegationTokenSecretManager", +"pipes Submitter", +"jobhistory JobUnsuccessfulCompletionEvent", +"event JobEvent", +"buffer OutputBuffer", +"event JobMapTaskRescheduledEvent", +"webapp AMWebApp", +"event TaskTAttemptFailedEvent", +"lib RegexMapper", +"split SplitMetaInfoReader", +"event TaskTAttemptEvent", +"pb RenewDelegationTokenRequestPBImpl", +"aggregate DoubleValueSum", +"mapreduce TaskAttemptContext", +"jobhistory TaskAttemptFinishedEvent", +"mapred YarnChild", +"client package-info", +"pb GetTaskAttemptReportResponsePBImpl", +"examples Join", +"mapreduce Counter", +"rm ContainerAllocator", +"pb KillJobRequestPBImpl", +"jobhistory NormalizedResourceEvent", +"aggregate ValueHistogram", +"jobhistory HistoryEvent", +"impl TaskImpl", +"hs UnparsedJob", +"serde NullWritableSerializer", +"checkpoint CheckpointID", +"aggregate ValueAggregatorJob", +"aggregate LongValueMin", +"filecache package-info", +"aggregate LongValueSum", +"mapred ReduceTask", +"mapreduce JobSubmissionFiles", +"serde VIntWritableSerializer", +"rm ContainerRequestEvent", +"webapp HsJobBlock", +"output SequenceFileAsBinaryOutputFormat", +"mapred FadvisedFileRegion", +"serde IntWritableSerializer", +"mapreduce JobResourceUploader", +"lib BinaryPartitioner", +"mapreduce TaskInputOutputContext", +"serde KVSerializer", +"mapreduce ClusterMetrics", +"event TaskRecoverEvent", +"records Counters", +"input SequenceFileAsTextInputFormat", +"db DBOutputFormat", +"speculate StartEndTimesBase", +"impl MapTaskImpl", +"lib CombineTextInputFormat", +"checkpoint CheckpointService", +"join TupleWritable", +"reduce Fetcher", +"buffer ByteBufferDataWriter", +"reduce MergeManager", +"launcher ContainerLauncherEvent", +"jobhistory JobHistoryUtils", +"aggregate ValueAggregator", +"db DataDrivenDBInputFormat", +"jobhistory TaskUpdatedEvent", +"join Parser", +"delegation DelegationTokenSelector", +"records JobState", +"aggregate UniqValueCount", +"mapred LocalContainerLauncher", +"terasort TeraOutputFormat", +"pb GetTaskReportResponsePBImpl", +"lib MultipleOutputFormat", +"pb CounterGroupPBImpl", +"math LongLong", +"join ComposableInputFormat", +"records TaskType", +"join WrappedRecordReader", +"mapred ShuffleHandler", +"webapp AMParams", +"mapred BasicTypeSorterBase", +"reduce ExceptionReporter", +"mapred TaskID", +"webapp NavBlock", +"db OracleDataDrivenDBRecordReader", +"util SizedWritable", +"mapreduce JobACL", +"hs HistoryServerStateStoreServiceFactory", +"examples Sort", +"local LocalContainerAllocator", +"hs JHSDelegationTokenSecretManager", +"lib TokenCountMapper", +"util MRProtoUtils", +"util BytesUtil", +"pipes BinaryProtocol", +"lib FieldSelectionMapReduce", +"mapred CleanupQueue", +"mapreduce Cluster", +"chain Chain", +"hs HistoryServerFileSystemStateStoreService", +"client ClientHSTokenSelector", +"mapred MapFileOutputFormat", +"mapred TIPStatus", +"input UncompressedSplitLineReader", +"protocol ClientProtocol", +"terasort TeraChecksum", +"delegation package-info", +"pb GetTaskReportsResponsePBImpl", +"mapred SequenceFileInputFormat", +"event JobCommitFailedEvent", +"lib IdentityMapper", +"mapred SequenceFileAsBinaryInputFormat", +"event TaskAttemptTooManyFetchFailureEvent", +"reduce MergeThread", +"mapred InputSplitWithLocationInfo", +"join InnerJoinRecordReader", +"terasort TeraSort", +"records AMInfo", +"lib MultipleOutputs", +"service package-info", +"partition TotalOrderPartitioner", +"webapp package-info", +"examples WordStandardDeviation", +"pb package-info", +"app TaskHeartbeatHandler", +"records TaskState", +"map TokenCounterMapper", +"jobtracker JTConfig", +"mapreduce QueueAclsInfo", +"pb CancelDelegationTokenResponsePBImpl", +"hs JobHistory", +"lib NullOutputFormat", +"app TaskAttemptFinishingMonitor", +"records JobReport", +"reduce MergeManagerImpl", +"records JobId", +"mapred FadvisedChunkedFile", +"pi Combinable", +"nativetask Constants", +"db DBSplitter", +"protocolrecords CancelDelegationTokenRequest", +"task TaskInputOutputContextImpl", +"uploader UploaderException", +"mapreduce InputFormat", +"lib InputSampler", +"terasort Unsigned16", +"aggregate ValueAggregatorCombiner", +"aggregate ValueAggregatorReducer", +"join CompositeRecordReader", +"counters AbstractCounterGroup", +"mapred PeriodicStatsAccumulator", +"dao JobsInfo", +"event JobSetupFailedEvent", +"job package-info", +"jobhistory JobPriorityChangeEvent", +"records CounterGroup", +"db IntegerSplitter", +"reduce ShuffleClientMetrics", +"speculate LegacyTaskRuntimeEstimator", +"map MultithreadedMapper", +"mapred Mapper", +"serde IKVSerializer", +"mapred TaskAttemptContextImpl", +"input TextInputFormat", +"mapred OutputCommitter", +"util HostUtil", +"mapred InputFormat", +"db DataDrivenDBRecordReader", +"protocolrecords CancelDelegationTokenResponse", +"pb FailTaskAttemptResponsePBImpl", +"mapred RunningJob", +"filecache package-info", +"nativetask DataReceiver", +"speculate DataStatistics", +"client ClientService", +"counters FileSystemCounterGroup", +"pb GetJobReportResponsePBImpl", +"mapred StatePeriodicStats", +"output MapFileOutputFormat", +"aggregate LongValueMax", +"mapreduce Mapper", +"mapred MergeSorter", +"pb GetJobReportRequestPBImpl", +"input MultipleInputs", +"pi DistBbp", +"output MultipleOutputs", +"jobhistory package-info", +"api MRDelegationTokenIdentifier", +"mapred TextInputFormat", +"job TaskAttemptStateInternal", +"terasort GenSort", +"fieldsel FieldSelectionMapper", +"dao AMAttemptsInfo", +"counters AbstractCounters", +"counters FrameworkCounterGroup", +"job TaskAttempt", +"mapred IndexCache", +"nativetask Command", +"jobhistory TaskFinishedEvent", +"split package-info", +"event JobFinishEvent", +"api HSAdminRefreshProtocol", +"input DelegatingMapper", +"join CompositeInputFormat", +"nativetask DataChannel", +"rm RMContainerRequestor", +"aggregate StringValueMin", +"mapred IFileOutputStream", +"protocolrecords KillTaskAttemptRequest", +"mapred LocalDistributedCacheManager", +"mapred JobConfigurable", +"hs HistoryStorage", +"counters LimitExceededException", +"webapp JobConfPage", +"token package-info", +"webapp HsJobsBlock", +"lib MultithreadedMapRunner", +"mapred FixedLengthInputFormat", +"event TaskAttemptContainerLaunchedEvent", +"job Task", +"lib HashPartitioner", +"math Modular", +"records TaskId", +"records TaskAttemptState", +"serde BoolWritableSerializer", +"mapreduce TaskReport", +"terasort TeraGen", +"mapred WrappedPeriodicStatsAccumulator", +"api HSAdminRefreshProtocolPB", +"hadoop CustomOutputCommitter", +"pi Parser", +"nativetask ICombineHandler", +"event TaskAttemptStatusUpdateEvent", +"mapred SortedRanges", +"input LineRecordReader", +"pb package-info", +"hs HistoryFileManager", +"reduce LongSumReducer", +"speculate TaskSpeculationPredicate", +"join ComposableRecordReader", +"examples Grep", +"protocolrecords GetJobReportRequest", +"jobhistory JobHistoryCopyService", +"mapred MapOutputFile", +"webapp HsController", +"job JobStateInternal", +"aggregate ValueAggregatorBaseDescriptor", +"join JoinRecordReader", +"mapreduce ContextFactory", +"counters GenericCounter", +"terasort TeraInputFormat", +"join OuterJoinRecordReader", +"counters CounterGroupBase", +"mapred TextOutputFormat", +"rm ResourceCalculatorUtils", +"mapred SplitLocationInfo", +"handlers NativeCollectorOnlyHandler", +"reduce InMemoryReader", +"mapred MapReduceBase", +"join TupleWritable", +"pb TaskAttemptIdPBImpl", +"preemption KillAMPreemptionPolicy", +"reduce WrappedReducer", +"mapreduce InputSplit", +"pb GetCountersRequestPBImpl", +"protocolrecords KillJobResponse", +"serde DoubleWritableSerializer", +"reduce ShuffleSchedulerImpl", +"examples ExampleDriver", +"mapred LocalJobRunnerMetrics", +"mapreduce Partitioner", +"pb GetTaskAttemptCompletionEventsResponsePBImpl", +"mapred JobContext", +"mapreduce OutputCommitter", +"pb KillTaskAttemptResponsePBImpl", +"mapred QueueConfigurationParser", +"mapreduce JobStatus", +"lib CombineSequenceFileInputFormat", +"mapred InputSplit", +"protocolrecords GetDiagnosticsResponse", +"protocolPB HSAdminRefreshProtocolClientSideTranslatorPB", +"dancing Pentomino", +"mapred IFile", +"db DBRecordReader", +"util package-info", +"mapred JVMId", +"hs HistoryServerStateStoreService", +"buffer InputBuffer", +"mapred DeprecatedQueueConfigurationParser", +"input FileSplit", +"mapred WrappedJvmID", +"mapred ShuffleConsumerPlugin", +"dao TaskAttemptInfo", +"aggregate StringValueMax", +"jobcontrol JobControl", +"join ResetableIterator", +"util NativeTaskOutput", +"protocolrecords GetTaskReportsResponse", +"output PartialFileOutputCommitter", +"records Phase", +"authorize package-info", +"mapred MultiFileSplit", +"pb TaskReportPBImpl", +"mapred FileAlreadyExistsException", +"pi TaskResult", +"protocolPB HSAdminRefreshProtocolServerSideTranslatorPB", +"impl ReduceTaskImpl", +"records TaskAttemptId", +"mapred Clock", +"lib IdentityReducer", +"mapreduce TaskID", +"jobcontrol Job", +"lib DelegatingInputFormat", +"db DBWritable", +"buffer BufferType", +"dancing OneSidedPentomino", +"mapred IndexRecord", +"mapred SequenceFileAsBinaryOutputFormat", +"reduce ShuffleHeader", +"speculate TaskRuntimeEstimator", +"impl JobImpl", +"nativetask CommandDispatcher", +"pb KillTaskRequestPBImpl", +"mapreduce Reducer", +"join CompositeInputSplit", +"math Montgomery", +"speculate NullTaskRuntimesEngine", +"mapred MapTask", +"webapp JAXBContextResolver", +"dancing DistributedPentomino", +"mapred JvmTask", +"app TaskAttemptListener", +"output PartialOutputCommitter", +"nativetask INativeHandler", +"lib FilterOutputFormat", +"mapred ReduceTaskAttemptImpl", +"aggregate UserDefinedValueAggregatorDescriptor", +"aggregate ValueHistogram", +"event TaskEvent", +"lib NLineInputFormat", +"mapreduce TaskCounter", +"mapred JobConf", +"serde BytesWritableSerializer", +"job Job", +"mapred TaskStatus", +"aggregate UniqValueCount", +"mapreduce StatusReporter", +"fieldsel FieldSelectionHelper", +"mapred IFileInputStream", +"mapreduce SharedCacheConfig", +"mapred FileSplit", +"jobhistory JobInitedEvent", +"pipes PipesReducer", +"event JobTaskAttemptCompletedEvent", +"pipes PipesNonJavaInputFormat", +"mapreduce JobCounter", +"handlers BufferPullee", +"mapred TaskLogAppender", +"launcher package-info", +"output PathOutputCommitterFactory", +"serde TextSerializer", +"dao MapTaskAttemptInfo", +"handlers BufferPuller", +"mapred TaskAttemptListenerImpl", +"reduce EventFetcher", +"pi Container", +"event JobTaskEvent", +"speculate package-info", +"hs CompletedTask", +"output FileOutputCommitterFactory", +"mapred Counters", +"dao AppInfo", +"reduce package-info", +"dao ReduceTaskAttemptInfo", +"join ResetableIterator", +"hs HSAuditLogger", +"db DBConfiguration", +"mapred WrappedProgressSplitsBlock", +"mapred Merger", +"jobhistory JobStatusChangedEvent", +"reduce Shuffle", +"terasort TeraScheduler", +"jobhistory AvroArrayUtils", +"output TextOutputFormat", +"webapp ConfBlock", +"aggregate ValueAggregatorJobBase", +"client HSAdmin", +"webapp HsConfPage", +"mapred NotRunningJob", +"mapred InvalidFileTypeException", +"pb CountersPBImpl", +"dao JobTaskAttemptCounterInfo", +"chain ChainReduceContextImpl", +"protocolrecords GetTaskReportRequest", +"util ResourceBundles", +"mapred QueueManager", +"mapred Reporter", +"nativetask Platforms", +"pi Util", +"jobhistory HistoryViewerPrinter", +"mapred BufferSorter", +"aggregate StringValueMax", +"input SplitLineReader", +"util LocalJobOutputFiles", +"client package-info", +"webapp HsWebServices", +"pb GetTaskAttemptReportRequestPBImpl", +"input CombineFileInputFormat", +"task package-info", +"webapp JobBlock", +"webapp HsTaskPage", +"dao HistoryInfo", +"mapreduce ID", +"mapred YARNRunner", +"examples QuasiMonteCarlo", +"service MRClientProtocolPBServiceImpl", +"input FixedLengthRecordReader", +"dao JobInfo", +"util MRApps", +"mapreduce CustomJobEndNotifier", +"aggregate ValueAggregatorDescriptor", +"webapp HsTasksPage", +"rm package-info", +"examples BaileyBorweinPlouffe", +"webapp HsWebApp", +"mapreduce TaskTrackerInfo", +"mapred SkipBadRecords", +"commit CommitterEvent", +"task MapContextImpl", +"protocolrecords package-info", +"hs CompletedJob", +"input SequenceFileAsTextRecordReader", +"mapred LocatedFileStatusFetcher", +"split JobSplit", +"reduce MapHost", +"input FixedLengthInputFormat", +"handlers BufferPushee", +"mapreduce Job", +"terasort TeraSortConfigKeys", +"records Counter", +"client package-info", +"speculate SimpleExponentialTaskRuntimeEstimator", +"mapreduce ReduceContext", +"handlers BufferPusher", +"protocolrecords GetJobReportResponse", +"lib MultipleSequenceFileOutputFormat", +"dao JobTaskAttemptState", +"db FloatSplitter", +"map InverseMapper", +"security TokenCache", +"util MRBuilderUtils", +"serde NativeSerialization", +"pi DistSum", +"input CombineFileRecordReader", +"webapp AttemptsPage", +"math Summation", +"jobhistory EventWriter", +"mapred YarnOutputFiles", +"hs package-info", +"event JobAbortCompletedEvent", +"checkpoint FSCheckpointID", +"pb JobReportPBImpl", +"db TextSplitter", +"webapp HsJobPage", +"examples AggregateWordHistogram", +"aggregate ValueAggregatorMapper", +"util NativeTaskOutputFiles", +"mapreduce QueueInfo", +"mapred MapTaskAttemptImpl", +"token JobTokenSelector", +"mapreduce OutputFormat", +"event TaskAttemptEventType", +"pb AMInfoPBImpl", +"examples RandomWriter", +"terasort TeraValidate", +"aggregate UserDefinedValueAggregatorDescriptor", +"reduce MapOutput", +"jobhistory TaskAttemptUnsuccessfulCompletionEvent", +"math Bellard", +"mapred ResourceMgrDelegate", +"output FilterOutputFormat", +"forecast package-info", +"serde ByteWritableSerializer", +"token JobTokenIdentifier", +"pipes Application", +"pipes UpwardProtocol", +"jobhistory JobHistoryEvent", +"pb JobIdPBImpl", +"serde INativeSerializer", +"client MRClientProtocolPBClientImpl", +"join CompositeRecordReader", +"protocolrecords KillTaskResponse", +"buffer DataInputStream", +"webapp SingleCounterBlock", +"delegation DelegationTokenIdentifier", +"output FileOutputFormatCounter", +"reduce OnDiskMapOutput", +"reduce InMemoryMapOutput", +"lib CombineFileInputFormat", +"protocolrecords GetTaskAttemptCompletionEventsResponse", +"protocol package-info", +"webapp TasksPage", +"aggregate ValueAggregatorDescriptor", +"mapred ClusterStatus", +"mapreduce TaskCompletionEvent", +"mapred ID", +"input SequenceFileInputFormat", +"aggregate ValueAggregatorBaseDescriptor", +"event TaskAttemptEvent", +"aggregate LongValueMin", +"hs HistoryServerNullStateStoreService", +"protocolrecords GetTaskAttemptReportRequest", +"nativetask NativeRuntime", +"lib TotalOrderPartitioner", +"event package-info", +"db DBInputFormat", +"filecache DistributedCache", +"output NullOutputFormat", +"reduce IntSumReducer", +"mapred JobInProgress", +"app ClusterInfo", +"db OracleDBRecordReader", +"aggregate LongValueSum", +"nativetask Platform", +"serde LongWritableSerializer", +"lib MultipleTextOutputFormat", +"aggregate StringValueMin", +"webapp InfoPage", +"db BigDecimalSplitter", +"webapp JAXBContextResolver", +"rm RMCommunicator", +"event TaskTAttemptKilledEvent", +"util OutputUtil", +"chain ChainMapContextImpl", +"webapp HsSingleCounterPage", +"event JobStartEvent", +"v2 LogParams", +"event JobSetupCompletedEvent", +"join CompositeInputSplit", +"serde FloatWritableSerializer", +"hs CachedHistoryStorage", +"protocolrecords GetCountersResponse", +"partition KeyFieldBasedComparator", +"launcher ContainerLauncher", +"security MRDelegationTokenRenewer", +"mapred JobQueueInfo", +"pipes PipesPartitioner", +"mapred RecordWriter", +"speculate Speculator", +"api MRClientProtocolPB", +"jobhistory TaskStartedEvent", +"preemption CheckpointAMPreemptionPolicy", +"hs CompletedTaskAttempt", +"buffer DataOutputStream", +"protocolrecords GetTaskAttemptCompletionEventsRequest", +"pb TaskAttemptCompletionEventPBImpl", +"filecache DistributedCache", +"event JobTaskAttemptFetchFailureEvent", +"uploader package-info", +"join OuterJoinRecordReader", +"protocolrecords GetDiagnosticsRequest", +"rm RMContainerAllocator", +"task JobContextImpl", +"mapreduce FileSystemCounter", +"security SecureShuffleUtils", +"mapreduce JobID", +"app MRAppMaster", +"mapred RawKeyValueIterator", +"webapp HsAboutPage", +"examples WordMean", +"records TaskReport", +"serde VLongWritableSerializer", +"db MySQLDBRecordReader", +"join StreamBackedIterator", +"nativetask NativeMapOutputCollectorDelegator", +"input NLineInputFormat", +"job TaskStateInternal", +"mapred MapRunnable", +"protocolrecords FailTaskAttemptResponse", +"mapred QueueRefresher", +"mapred StatisticsCollector", +"mapreduce RecordWriter", +"commit CommitterEventHandler", +"mapred KeyValueLineRecordReader", +"app JobEndNotifier", +"jobhistory MapAttemptFinishedEvent", +"output FileOutputCommitter", +"aggregate LongValueMax", +"mapred JobProfile", +"mapred MapReduceChildJVM", +"aggregate ValueAggregatorReducer", +"input DelegatingInputFormat", +"join WrappedRecordReader", +"lib LazyOutputFormat", +"hs HistoryContext", +"db DBOutputFormat", +"dao JobsInfo", +"event JobCommitCompletedEvent", +"mapred MultiFileInputFormat", +"mapred TaskUmbilicalProtocol", +"nativetask INativeComparable", +"mapred Task", +"partition KeyFieldHelper", +"hs JobHistoryServer", +"pi SummationWritable", +"buffer ByteBufferDataReader", +"mapred KeyValueTextInputFormat", +"api HSAdminProtocol", +"protocol ClientProtocolProvider", +"join ArrayListBackedIterator", +"input SequenceFileAsBinaryInputFormat", +"lib KeyFieldBasedComparator", +"input CombineFileRecordReaderWrapper", +"nativetask HadoopPlatform", +"speculate SpeculatorEvent", +"speculate DefaultSpeculator", +"webapp AppView", +"webapp HsNavBlock", +"jobhistory HistoryViewer", +"mapreduce JobPriority", +"filecache ClientDistributedCacheManager", +"jobcontrol JobControl", +"mapred YarnClientProtocolProvider", +"commit CommitterJobAbortEvent", +"api MRClientProtocol", +"reduce InMemoryWriter", +"util MRWebAppUtil", +"proto HSClientProtocol", +"join Parser", +"mapreduce MRConfig", +"join ComposableInputFormat", +"reduce ShuffleScheduler", +"examples DBCountPageView", +"forecast SimpleExponentialSmoothing", +"join ArrayListBackedIterator", +"mapred JobPriority", +"util ConfigUtil", +"mapreduce JobContext", +"db OracleDataDrivenDBInputFormat", +"chain ChainReducer", +"db MySQLDataDrivenDBRecordReader", +"mapred SequenceFileAsTextRecordReader", +"lib ChainReducer", +"partition HashPartitioner", +"lib MultipleInputs", +"mapred SequenceFileOutputFormat", +"webapp CountersPage", +"jobhistory JobSummary", +"mapred ReduceTaskStatus", +"db DateSplitter", +"protocolrecords KillJobRequest", +"mapreduce TaskType", +"records Locality", +"metrics package-info", +"mapred TaskAttemptID", +"webapp MapReduceTrackingUriPlugin", +"mapred OutputFormat", +"mapred Master", +"webapp HsAttemptsPage", +"launcher ContainerLauncherImpl", +"mapred FileInputFormat", +"mapred SpillRecord", +"mapreduce CryptoUtils", +"pipes PipesMapRunner", +"event TaskAttemptFailEvent", +"records TaskAttemptCompletionEventStatus", +"counters CounterGroupFactory", +"commit package-info", +"mapred InvalidInputException", +"webapp AMWebServices", +"pb GetDelegationTokenResponsePBImpl", +"api package-info", +"mapreduce MarkableIterator", +"app MRClientSecurityInfo", +"join ComposableRecordReader", +"counters Limits", +"token JobTokenSecretManager", +"commit CommitterJobCommitEvent", +"mapred LocalJobRunner", +"rm ContainerFailedEvent", +"event JobEventType", +"partition BinaryPartitioner", +"webapp AppController", +"handlers IDataLoader", +"mapred SequenceFileAsTextInputFormat", +"webapp HsTasksBlock", +"mapred ClientServiceDelegate", +"client MRClientService", +"nativetask TaskContext", +"pb KillTaskResponsePBImpl", +"input FileInputFormat", +"join MultiFilterRecordReader", +"mapred FileOutputCommitter", +"mapred LineRecordReader", +"api HSClientProtocolPB", +"lib DelegatingMapper", +"mapred SequenceFileInputFilter", +"lib ChainMapper", +"jobcontrol ControlledJob", +"dao TaskCounterInfo", +"protocolrecords GetTaskReportsRequest", +"util package-info", +"math ArithmeticProgression", +"dao TaskInfo", +"app package-info", +"records TaskAttemptCompletionEvent", +"pb CancelDelegationTokenRequestPBImpl", +"checkpoint CheckpointNamingService", +"server HSAdminServer", +"aggregate ValueAggregatorJob", +"pb GetDiagnosticsRequestPBImpl", +"webapp CountersBlock", +"serde DefaultSerializer", +"webapp App", +"mapred MROutputFiles", +"mapreduce MarkableIteratorInterface", +"client ClientHSSecurityInfo", +"uploader DefaultJars", +"tasktracker TTConfig", +"lib CombineFileRecordReaderWrapper", +"lib LongSumReducer", +"mapred RamManager", +"input CombineFileSplit", +"mapred Partitioner", +"protocolrecords GetCountersRequest", +"aggregate ValueAggregatorJobBase", +"impl package-info", +"records package-info", +"input KeyValueLineRecordReader", +"mapred QueueACL", +"event JobCounterUpdateEvent", +"pb GetTaskAttemptCompletionEventsRequestPBImpl", +"util LocalResourceBuilder", +"mapred RecordReader", +"jobhistory JSONHistoryViewerPrinter", +"mapred FileOutputFormat", +"jobhistory JobQueueChangeEvent", +"output NamedCommitterFactory", +"checkpoint TaskCheckpointID", +"pb KillTaskAttemptRequestPBImpl", +"counters package-info", +"db OracleDateSplitter", +"mapred AMFeedback", +"aggregate ValueAggregator", +"dao TaskCounterGroupInfo", +"protocolrecords FailTaskAttemptRequest", +"handlers CombinerHandler", +"jobhistory JobHistoryParser", +"util ReadWriteBuffer", +"examples WordCount", +"mapred MRConstants", +"rm RMContainerAllocationException", +"join CompositeInputFormat", +"join StreamBackedIterator", +"pipes OutputHandler", +"preemption AMPreemptionPolicy", +"reduce IFileWrappedMapOutput", +"uploader FrameworkUploader", +"dao TasksInfo", +"examples SecondarySort", +"event TaskEventType", +"input KeyValueTextInputFormat", +"mapred ClientCache", +"input SequenceFileRecordReader", +"join OverrideRecordReader", +"nativetask NativeDataSource", +"jobhistory JHAdminConfig", +"nativetask NativeBatchProcessor", +"dao JobTaskCounterInfo", +"mapred Queue", +"pb CounterPBImpl", +"mapred JobID", +"launcher ContainerRemoteLaunchEvent", +"split JobSplitWriter", +"jobhistory JobInfoChangeEvent", +"fieldsel FieldSelectionReducer", +"service HSClientProtocolPBServiceImpl", +"mapred Reducer", +"pb GetTaskReportRequestPBImpl", +"protocolrecords GetTaskReportResponse", +"util MRJobConfUtil", +"jobhistory FileNameIndexUtils", +"aggregate ValueAggregatorMapper", +"partition RehashPartitioner", +"jobhistory package-info", +"join JoinRecordReader", +"protocolrecords GetTaskAttemptReportResponse", +"mapred CumulativePeriodicStats", +"pb KillJobResponsePBImpl", +"task TaskAttemptContextImpl", +"pb GetTaskReportsRequestPBImpl", +"reduce LocalFetcher", +"partition InputSampler", +"db BooleanSplitter", +"join MultiFilterRecordReader", +"dao ConfInfo", +"lib Chain", +"event TaskAttemptContainerAssignedEvent", +"task ReduceContextImpl", +"input CombineSequenceFileInputFormat", +"mapreduce QueueState", +"mapred JobACLsManager", +"mapred JobStatus", +"protocolrecords GetDelegationTokenResponse", +"jobhistory EventReader", +"mapred TaskAttemptContext", +"nativetask NativeDataTarget", +"event JobUpdatedNodesEvent", +"event TaskAttemptKillEvent", +"mapred OutputLogFilter", +"map RegexMapper", +"event TaskAttemptRecoverEvent", +"pb TaskAttemptReportPBImpl", +"pb FailTaskAttemptRequestPBImpl", +"metrics MRAppMetrics", +"aggregate DoubleValueSum", +"impl TaskAttemptImpl", +"jobhistory JobSubmittedEvent", +"dao AMAttemptsInfo", +"commit CommitterTaskAbortEvent", +"app AppContext", +"mapred JobClient", +"input DelegatingRecordReader", +"output PathOutputCommitter", +"mapred JobInfo", +"hs HistoryClientService", +"counters AbstractCounter", +"mapred Utils", +"serde SerializationFramework", +"mapreduce MRJobConfig", +"mapred TaskCompletionEvent", +"hs PartialJob", +"records TaskAttemptReport", +"terasort Random16", +"dao CounterGroupInfo", +"commit CommitterJobSetupEvent", +"records Avataar", +"webapp JobPage", +"mapreduce Counters", +"mapreduce TaskAttemptID", +"mapred InvalidJobConfException", +"examples AggregateWordCount", +"pb GetCountersResponsePBImpl", +"mapred SequenceFileRecordReader", +"dao AMAttemptInfo", +"dao AMAttemptInfo", +"checkpoint SimpleNamingService", +"output SequenceFileOutputFormat", +"dao JobCounterInfo", +"input CombineTextInputFormat", +"protocolrecords RenewDelegationTokenRequest", +"dancing Sudoku", +"dao BlacklistedNodesInfo", +"mapred TaskReport", +"event JobDiagnosticsUpdateEvent", +"hs HistoryServerLeveldbStateStoreService", +"mapreduce MapContext", +"checkpoint RandomNameCNS", +"protocolrecords GetDelegationTokenRequest", +"mapred QueueAclsInfo", +"mapred MapTaskCompletionEventsUpdate", +"util JobHistoryEventUtils", +"tools CLI", +"client HSClientProtocolPBClientImpl", +"mapred LocalClientProtocolProvider", +"db DBInputFormat", +"webapp SingleCounterPage", +"jobhistory TaskFailedEvent", +"examples WordMedian", +"dao JobInfo", +"jobhistory JobIndexInfo", +"output FileOutputFormat", +"authorize ClientHSPolicyProvider", +"pb GetDiagnosticsResponsePBImpl", +"mapred JobContextImpl", +"examples RandomTextWriter", +"input TaggedInputSplit", +"checkpoint EnumCounter", +"mapreduce TypeConverter", +"nativetask StatusReportChecker", +"webapp HsLogsPage", +"webapp TasksBlock", +"aggregate ValueAggregatorCombiner", +"input CompressedSplitLineReader", +"examples MultiFileWordCount", +"webapp JobsBlock", +"jobhistory TaskAttemptStartedEvent", +"jobhistory JobFinishedEvent", +"webapp TaskPage", +"mapred JobQueueClient", +"protocolrecords KillTaskAttemptResponse", +"local package-info", +"mapred JobEndNotifier", +"mapred BackupStore", +"lib CombineFileRecordReader", +"protocolrecords KillTaskRequest", +"pipes DownwardProtocol", +"mapred OutputCollector", +"checkpoint FSCheckpointService", +"join OverrideRecordReader", +"lib InverseMapper", +"input FileInputFormatCounter", +"mapred Operation", +"mapred FixedLengthRecordReader", +"webapp package-info", +"util ProcessTree", +"jobhistory ReduceAttemptFinishedEvent", +"db DBWritable", +"pb TaskIdPBImpl", +"dao CounterInfo", +"mapred TaskLog", +"lib TaggedInputSplit", +"rm ContainerAllocatorEvent", +"protocolrecords RenewDelegationTokenResponse", +"partition KeyFieldBasedPartitioner", +"mapred MapRunner", +"join InnerJoinRecordReader", +"speculate ExponentiallySmoothedTaskRuntimeEstimator", +"webapp HsCountersPage", +"mapreduce RecordReader", +"lib KeyFieldBasedPartitioner", +"map WrappedMapper", +"pb RenewDelegationTokenResponsePBImpl", +"dancing DancingLinks", +"dao ConfEntryInfo" \ No newline at end of file diff --git a/src/main/resources/cdtocode/Hadoop MapReduce-VSMLSI.xls b/src/main/resources/cdtocode/Hadoop MapReduce-VSMLSI.xls new file mode 100644 index 0000000000000000000000000000000000000000..39fc5e40559ec71c24f71d4bc40c4556fc5655d0 Binary files /dev/null and b/src/main/resources/cdtocode/Hadoop MapReduce-VSMLSI.xls differ diff --git a/src/main/resources/cdtocode/Hadoop MapReduce-VSMLSIdata.xls b/src/main/resources/cdtocode/Hadoop MapReduce-VSMLSIdata.xls new file mode 100644 index 0000000000000000000000000000000000000000..bda14a502c7130f5e1a06afc1c7737518b1eb434 Binary files /dev/null and b/src/main/resources/cdtocode/Hadoop MapReduce-VSMLSIdata.xls differ diff --git a/src/main/resources/cdtocode/Hadoop MapReduce-nodoc.xls b/src/main/resources/cdtocode/Hadoop MapReduce-nodoc.xls new file mode 100644 index 0000000000000000000000000000000000000000..332db46db86a6d7d2e803a36d6026ee1caacc0f4 Binary files /dev/null and b/src/main/resources/cdtocode/Hadoop MapReduce-nodoc.xls differ diff --git a/src/main/resources/cdtocode/Hadoop MapReduce-nolog.xls b/src/main/resources/cdtocode/Hadoop MapReduce-nolog.xls new file mode 100644 index 0000000000000000000000000000000000000000..9ac762ce4f7ee40400f15ea4f57393a8418802e1 Binary files /dev/null and b/src/main/resources/cdtocode/Hadoop MapReduce-nolog.xls differ diff --git a/src/main/resources/cdtocode/Hadoop MapReduce-vsmlsiraw.xls b/src/main/resources/cdtocode/Hadoop MapReduce-vsmlsiraw.xls new file mode 100644 index 0000000000000000000000000000000000000000..3dbe7f22243c18d049e062acf05e108153c33164 Binary files /dev/null and b/src/main/resources/cdtocode/Hadoop MapReduce-vsmlsiraw.xls differ diff --git a/src/main/resources/cdtocode/cd/cd-Hadoop HDFS1.png b/src/main/resources/cdtocode/cd/cd-Hadoop HDFS1.png new file mode 100644 index 0000000000000000000000000000000000000000..cb7e560e8df5ddc227ccd75b086fc4729cbe4ec3 Binary files /dev/null and b/src/main/resources/cdtocode/cd/cd-Hadoop HDFS1.png differ diff --git a/src/main/resources/cdtocode/cd/cd-Hadoop HDFS2.png b/src/main/resources/cdtocode/cd/cd-Hadoop HDFS2.png new file mode 100644 index 0000000000000000000000000000000000000000..6a89ca7fd827bc457878f9788eab892f1fae1a76 Binary files /dev/null and b/src/main/resources/cdtocode/cd/cd-Hadoop HDFS2.png differ diff --git a/src/main/resources/cdtocode/cd/cd-Hadoop MapReduce1.png b/src/main/resources/cdtocode/cd/cd-Hadoop MapReduce1.png new file mode 100644 index 0000000000000000000000000000000000000000..9a7cfa501497a3917a17a75fd45e5497e93e1593 Binary files /dev/null and b/src/main/resources/cdtocode/cd/cd-Hadoop MapReduce1.png differ diff --git a/src/main/resources/cdtocode/cd/cd-Hadoop MapReduce2.png b/src/main/resources/cdtocode/cd/cd-Hadoop MapReduce2.png new file mode 100644 index 0000000000000000000000000000000000000000..ffc790c72b0e6e6ffeea92cb8fd31c857981dffb Binary files /dev/null and b/src/main/resources/cdtocode/cd/cd-Hadoop MapReduce2.png differ diff --git a/src/main/resources/cdtocode/doc/Apache OODT File Manager/A FRAMEWORK FOR COLLABORATIVE REVIEW OF CANDIDATE EVENTS IN HIGH DATA RATE STREAMS THE V-FASTR EXPERIMENT AS A CASE STUDY.txt.xml.xls b/src/main/resources/cdtocode/doc/Apache OODT File Manager/A FRAMEWORK FOR COLLABORATIVE REVIEW OF CANDIDATE EVENTS IN HIGH DATA RATE STREAMS THE V-FASTR EXPERIMENT AS A CASE STUDY.txt.xml.xls index 6a7e54ee95a8fecab006df314349ac4b66ff9f8e..00500ea9ef89f7c35d8d152086a7cd8b74a0c1a7 100644 Binary files a/src/main/resources/cdtocode/doc/Apache OODT File Manager/A FRAMEWORK FOR COLLABORATIVE REVIEW OF CANDIDATE EVENTS IN HIGH DATA RATE STREAMS THE V-FASTR EXPERIMENT AS A CASE STUDY.txt.xml.xls and b/src/main/resources/cdtocode/doc/Apache OODT File Manager/A FRAMEWORK FOR COLLABORATIVE REVIEW OF CANDIDATE EVENTS IN HIGH DATA RATE STREAMS THE V-FASTR EXPERIMENT AS A CASE STUDY.txt.xml.xls differ diff --git a/src/test/java/com/hy/java/uct/cdtocode/DocAnalyzerTest.java b/src/test/java/com/hy/java/uct/cdtocode/DocAnalyzerTest.java index a3299cb666849bb9dee9eb294477468c0d27b455..24a52225ddd16081f762f7a110d90de1a5519018 100644 --- a/src/test/java/com/hy/java/uct/cdtocode/DocAnalyzerTest.java +++ b/src/test/java/com/hy/java/uct/cdtocode/DocAnalyzerTest.java @@ -72,13 +72,15 @@ public class DocAnalyzerTest { Pattern pattern = Pattern.compile("([a-zA-Z_$][a-zA-Z\\d_$]*\\.)*[a-zA-Z_$][a-zA-Z\\d_$]*", Pattern.CASE_INSENSITIVE); Matcher matcher = null; // 检测 - matcher = pattern.matcher(":oejs.AbstractConnector:main:"); - ":oejs.AbstractConnector:main:".matches("([a-zA-Z_$][a-zA-Z\\d_$]*\\.)*[a-zA-Z_$][a-zA-Z\\d_$]*"); + // matcher = pattern.matcher(":oejs.AbstractConnector:main:"); + matcher = pattern.matcher("java.util.ArrayList"); + "java.util.ArrayList".matches("([a-zA-Z_$][a-zA-Z\\d_$]*\\.)*[a-zA-Z_$][a-zA-Z\\d_$]*"); while (matcher.find()) { res = true; break; } - System.out.println(":oejs.AbstractConnector:main:".matches("([a-zA-Z_$][a-zA-Z\\d_$]*\\.)*[a-zA-Z_$][a-zA-Z\\d_$]*")); + System.out.println(res); + System.out.println("java.util.ArrayList".matches("([a-zA-Z_$][a-zA-Z\\d_$]*\\.)*[a-zA-Z_$][a-zA-Z\\d_$]*")); } /*