prefix
stringlengths
82
32.6k
middle
stringlengths
5
470
suffix
stringlengths
0
81.2k
file_path
stringlengths
6
168
repo_name
stringlengths
16
77
context
listlengths
5
5
lang
stringclasses
4 values
ground_truth
stringlengths
5
470
/* * Created on Sun Mar 26 2023 * * Copyright (c) storycraft. Licensed under the Apache Licence 2.0. */ package sh.pancake.serdemc.data.nbt.io; import java.io.IOException; import java.util.Map; import javax.annotation.Nullable; import lombok.AllArgsConstructor; import sh.pancake.serdemc.data.nbt.NbtRootCompound; import sh.pancake.serdemc.data.nbt.NbtTagList; import sh.pancake.serdemc.data.nbt.NbtTagValue; import sh.pancake.serdemc.io.DataWriter; @AllArgsConstructor public class NbtWriter { private final DataWriter writer; public void writeByte(byte value) throws IOException { writer.writeByte(value); } public void writeShort(short value) throws IOException { writer.writeShort(value); } public void writeInt(int value) throws IOException { writer.writeInt(value); } public void writeLong(long value) throws IOException { writer.writeLong(value); } public void writeFloat(float value) throws IOException { writer.writeFloat(value); } public void writeDouble(double value) throws IOException { writer.writeDouble(value); } public void writeByteArray(byte[] value) throws IOException {
writer.writeInt(value.length);
writer.writeBytes(value); } public void writeString(String value) throws IOException { writer.writeShortUnsigned(value.length()); writer.writeBytes(value.getBytes()); } public <Item> void writeList(NbtTagList<Item> list) throws IOException { writer.writeByte(list.getType()); writer.writeInt(list.size()); for (NbtTagValue<?> tag : list) { writeTag(tag); } } public void writeCompound(Map<String, NbtTagValue<?>> compound) throws IOException { for (String name : compound.keySet()) { NbtTagValue<?> tag = compound.get(name); writer.writeByte(tag.getType()); writeString(name); writeTag(tag); } writer.writeByte(NbtTagValue.TAG_END); } public void writeRootCompoundOptional(@Nullable NbtRootCompound compound) throws IOException { if (compound == null) { writeByte(NbtTagValue.TAG_END); return; } writeRootCompound(compound); } public void writeRootCompound(NbtRootCompound compound) throws IOException { writeByte(NbtTagValue.TAG_COMPOUND); writeString(compound.getName()); writeCompound(compound.getCompound()); } public void writeIntArray(int[] value) throws IOException { int length = value.length; writer.writeInt(length); for (int i = 0; i < length; i++) { writeInt(value[i]); } } public void writeLongArray(long[] value) throws IOException { int length = value.length; writer.writeInt(length); for (int i = 0; i < length; i++) { writeLong(value[i]); } } @SuppressWarnings("unchecked") public void writeTag(NbtTagValue<?> tag) throws IOException { byte type = tag.getType(); Object value = tag.getValue(); switch (type) { case NbtTagValue.TAG_BYTE: { writeByte((byte) value); return; } case NbtTagValue.TAG_SHORT: { writeShort((short) value); return; } case NbtTagValue.TAG_INT: { writeInt((int) value); return; } case NbtTagValue.TAG_LONG: { writeLong((long) value); return; } case NbtTagValue.TAG_FLOAT: { writeFloat((float) value); return; } case NbtTagValue.TAG_DOUBLE: { writeDouble((double) value); return; } case NbtTagValue.TAG_BYTE_ARRAY: { writeByteArray((byte[]) value); return; } case NbtTagValue.TAG_STRING: { writeString((String) value); return; } case NbtTagValue.TAG_LIST: { writeList((NbtTagList<?>) value); return; } case NbtTagValue.TAG_COMPOUND: { writeCompound((Map<String, NbtTagValue<?>>) value); return; } case NbtTagValue.TAG_INT_ARRAY: { writeIntArray((int[]) value); return; } case NbtTagValue.TAG_LONG_ARRAY: { writeLongArray((long[]) value); return; } default: throw new RuntimeException("Unknown NBT type: " + type); } } }
api/src/main/java/sh/pancake/serdemc/data/nbt/io/NbtWriter.java
storycraft-serde-mc-757a9e4
[ { "filename": "api/src/main/java/sh/pancake/serdemc/io/OutputStreamWriter.java", "retrieved_chunk": " @Override\n public void writeShort(short value) throws IOException {\n stream.writeShort(value);\n }\n @Override\n public void writeInt(int value) throws IOException {\n stream.writeInt(value);\n }\n @Override\n public void writeLong(long value) throws IOException {", "score": 0.944202184677124 }, { "filename": "api/src/main/java/sh/pancake/serdemc/io/DataWriter.java", "retrieved_chunk": " default void writeShortUnsigned(int value) throws IOException {\n writeShort((short) value);\n }\n void writeInt(int value) throws IOException;\n void writeLong(long value) throws IOException;\n default void writeFloat(float value) throws IOException {\n writeInt(Float.floatToRawIntBits(value));\n }\n default void writeDouble(double value) throws IOException {\n writeLong(Double.doubleToRawLongBits(value));", "score": 0.94209223985672 }, { "filename": "api/src/main/java/sh/pancake/serdemc/network/io/PacketDataWriter.java", "retrieved_chunk": " public void writeBoolean(boolean value) throws IOException {\n writer.writeByte((byte) (value ? 1 : 0));\n }\n public void writeBytes(byte[] dst) throws IOException {\n writer.writeBytes(dst);\n }\n public void writeShort(short value) throws IOException {\n writer.writeShort(value);\n }\n public void writeShortUnsigned(int value) throws IOException {", "score": 0.939257800579071 }, { "filename": "api/src/main/java/sh/pancake/serdemc/network/io/PacketDataWriter.java", "retrieved_chunk": " writer.writeShortUnsigned(value);\n }\n public void writeInt(int value) throws IOException {\n writer.writeInt(value);\n }\n public void writeVarInt(int value) throws IOException {\n while (true) {\n if ((value & ~0x7F) == 0) {\n writeByte((byte) value);\n return;", "score": 0.9356204867362976 }, { "filename": "api/src/main/java/sh/pancake/serdemc/network/io/PacketDataWriter.java", "retrieved_chunk": " if ((value & ~0x7FL) == 0) {\n writeByte((byte) value);\n return;\n }\n writeByte((byte) ((value & 0x7F) | 0x80));\n value >>>= 7;\n }\n }\n public void writeFloat(float value) throws IOException {\n writer.writeFloat(value);", "score": 0.9266714453697205 } ]
java
writer.writeInt(value.length);
package top.sssd.ddns.service.impl; import com.aliyun.alidns20150109.Client; import com.aliyun.alidns20150109.models.DescribeSubDomainRecordsResponse; import com.aliyun.alidns20150109.models.DescribeSubDomainRecordsResponseBody; import lombok.extern.slf4j.Slf4j; import org.springframework.http.HttpStatus; import org.springframework.stereotype.Service; import org.springframework.util.StringUtils; import top.sssd.ddns.common.BizException; import top.sssd.ddns.common.enums.RecordTypeEnum; import top.sssd.ddns.common.utils.DoMainUtil; import top.sssd.ddns.model.entity.ParsingRecord; import top.sssd.ddns.service.DynamicDnsService; import top.sssd.ddns.utils.AliDnsUtils; /** * @author sssd * @careate 2023-03-20-13:41 */ @Service @Slf4j public class AliDynamicDnsServiceImpl implements DynamicDnsService { @Override public boolean exist(String serviceProviderId, String serviceProviderSecret, String subDomain, String recordType) throws Exception { Client client = AliDnsUtils.createClient(serviceProviderId, serviceProviderSecret); DescribeSubDomainRecordsResponse response = null; try { response = AliDnsUtils.getSubDomainParseList(client, subDomain, recordType); } catch (Exception e) { return false; } if (response.statusCode != HttpStatus.OK.value()) { log.error("调用阿里云DNS解析失败,请检查传入的serviceProviderId,serviceProviderSecret,域名是否正确"); throw new BizException("调用阿里云DNS解析失败,请检查传入的serviceProviderId,serviceProviderSecret,域名是否正确"); } DescribeSubDomainRecordsResponseBody body = response.getBody(); return body.getTotalCount() > 0 ; } @Override public void add(ParsingRecord parsingRecord, String ip) throws Exception { //call dns api Client client = AliDnsUtils.createClient(parsingRecord.getServiceProviderId(), parsingRecord.getServiceProviderSecret()); String subDoMain = parsingRecord.getDomain(); String domain = null; String rr = null; if (
DoMainUtil.firstLevel(subDoMain)) {
domain = subDoMain; rr = "@"; } else { domain = subDoMain.substring(findNthOccurrence(subDoMain, ".", 1) + 1); rr = subDoMain.substring(0, findNthOccurrence(subDoMain, ".", 1)); } AliDnsUtils.add(client, domain, rr, RecordTypeEnum.getNameByIndex(parsingRecord.getRecordType()), ip); } public static int findNthOccurrence(String str, String subStr, int n) { // 记录出现次数 int count = 0; // 从后往前查找最后一次出现的位置 int index = str.lastIndexOf(subStr); // 如果找到了并且出现次数小于n while (index != -1 && count < n) { // 继续往前查找下一次出现的位置 index = str.lastIndexOf(subStr, index - 1); // 更新出现次数 count++; } // 返回最后一次出现的位置的索引 return index; } @Override public void update(ParsingRecord parsingRecord, String ip,String recordId) throws Exception { //call dns api Client client = AliDnsUtils.createClient(parsingRecord.getServiceProviderId(), parsingRecord.getServiceProviderSecret()); String subDoMain = parsingRecord.getDomain(); String rr = null; if (DoMainUtil.firstLevel(subDoMain)) { rr = "@"; } else { rr = subDoMain.substring(0, findNthOccurrence(subDoMain, ".", 1)); } String recordTypeName = RecordTypeEnum.getNameByIndex(parsingRecord.getRecordType()); AliDnsUtils.update(client, recordId, rr, recordTypeName, ip); } @Override public String getRecordId(ParsingRecord parsingRecord, String ip) throws Exception { //call dns api Client client = AliDnsUtils.createClient(parsingRecord.getServiceProviderId(), parsingRecord.getServiceProviderSecret()); String subDoMain = parsingRecord.getDomain(); String recordTypeName = RecordTypeEnum.getNameByIndex(parsingRecord.getRecordType()); String recordId = AliDnsUtils.getDomainRecordId(client, subDoMain, recordTypeName, ip);; if (StringUtils.isEmpty(recordId)) { throw new BizException("没有该域名对应的解析记录"); } return recordId; } @Override public void remove(ParsingRecord parsingRecord, String ip) throws Exception { Client client = AliDnsUtils.createClient(parsingRecord.getServiceProviderId(), parsingRecord.getServiceProviderSecret()); String recordTypeName = RecordTypeEnum.getNameByIndex(parsingRecord.getRecordType()); String recordId = AliDnsUtils.getDomainRecordId(client, parsingRecord.getDomain(), recordTypeName, ip); AliDnsUtils.delete(client, recordId); } @Override public String getIpBySubDomainWithType(ParsingRecord parsingRecord) throws Exception { Client client = AliDnsUtils.createClient(parsingRecord.getServiceProviderId(), parsingRecord.getServiceProviderSecret()); return AliDnsUtils.getIpBySubDomainWithType(client, parsingRecord.getDomain(), RecordTypeEnum.getNameByIndex(parsingRecord.getRecordType())); } }
src/main/java/top/sssd/ddns/service/impl/AliDynamicDnsServiceImpl.java
sssdgithub-ddns4j-eadfbee
[ { "filename": "src/main/java/top/sssd/ddns/service/impl/TencentDynamicDnsServiceImpl.java", "retrieved_chunk": " public void add(ParsingRecord parsingRecord, String ip) throws TencentCloudSDKException {\n String domain = parsingRecord.getDomain();\n String resultDomain = \"\";\n String subDoMain = \"\";\n if (DoMainUtil.firstLevel(domain)) {\n subDoMain = \"@\";\n } else {\n resultDomain = domain.substring(domain.indexOf('.') + 1);\n subDoMain = domain.substring(0, domain.indexOf('.'));\n }", "score": 0.9374330639839172 }, { "filename": "src/main/java/top/sssd/ddns/service/impl/TencentDynamicDnsServiceImpl.java", "retrieved_chunk": " TencentDnsUtils.createRecord(resultDomain, subDoMain, RecordTypeEnum.getNameByIndex(parsingRecord.getRecordType()), parsingRecord.getServiceProviderId(), parsingRecord.getServiceProviderSecret(), ip);\n }\n @Override\n public void update(ParsingRecord parsingRecord, String ip, String recordId) {\n String domain = parsingRecord.getDomain();\n String resultDomain = \"\";\n String subDoMain = \"\";\n if (DoMainUtil.firstLevel(domain)) {\n subDoMain = \"@\";\n } else {", "score": 0.9371177554130554 }, { "filename": "src/main/java/top/sssd/ddns/service/impl/CloudflareDynamicDnsServiceImpl.java", "retrieved_chunk": " }\n @Override\n public void add(ParsingRecord parsingRecord, String ip) throws JsonProcessingException {\n CloudflareUtils.CloudflareResponse cloudflareResponse =\n CloudflareUtils.add(parsingRecord.getServiceProviderId(), parsingRecord.getServiceProviderSecret(), parsingRecord.getDomain(), RecordTypeEnum.getNameByIndex(parsingRecord.getRecordType()), ip);\n if (Boolean.FALSE.equals(cloudflareResponse.getSuccess())) {\n log.error(\"域名添加失败:{}\", parsingRecord);\n throw new BizException(\"域名添加失败\");\n }\n }", "score": 0.9215712547302246 }, { "filename": "src/main/java/top/sssd/ddns/service/impl/ParsingRecordServiceImpl.java", "retrieved_chunk": " @Override\n public void add(ParsingRecord parsingRecord) throws Exception {\n DynamicDnsService dynamicDnsService = DynamicDnsServiceFactory.getServiceInstance(parsingRecord.getServiceProvider());\n String ip = getIp(parsingRecord);\n //后端唯一性校验\n ParsingRecord checkParsingRecord = this.lambdaQuery()\n .eq(ParsingRecord::getServiceProvider, parsingRecord.getServiceProvider())\n .eq(ParsingRecord::getRecordType, parsingRecord.getRecordType())\n .eq(ParsingRecord::getDomain, parsingRecord.getDomain())\n .eq(ParsingRecord::getIp, parsingRecord.getIp())", "score": 0.9164733290672302 }, { "filename": "src/main/java/top/sssd/ddns/service/impl/TencentDynamicDnsServiceImpl.java", "retrieved_chunk": " resultDomain = domain.substring(domain.indexOf('.') + 1);\n subDoMain = domain.substring(0, domain.indexOf('.'));\n }\n try {\n TencentDnsUtils.updateRecord(resultDomain, subDoMain, RecordTypeEnum.getNameByIndex(parsingRecord.getRecordType()), parsingRecord.getServiceProviderId(), parsingRecord.getServiceProviderSecret(), ip, Long.parseLong(recordId));\n } catch (TencentCloudSDKException e) {\n e.printStackTrace();\n }\n }\n @Override", "score": 0.9163144826889038 } ]
java
DoMainUtil.firstLevel(subDoMain)) {
/* * Created on Sun Mar 26 2023 * * Copyright (c) storycraft. Licensed under the Apache Licence 2.0. */ package sh.pancake.serdemc.data.nbt.io; import java.io.IOException; import java.nio.charset.StandardCharsets; import javax.annotation.Nullable; import lombok.AllArgsConstructor; import sh.pancake.serdemc.data.nbt.NbtRootCompound; import sh.pancake.serdemc.data.nbt.NbtTagCompound; import sh.pancake.serdemc.data.nbt.NbtTagList; import sh.pancake.serdemc.data.nbt.NbtTagValue; import sh.pancake.serdemc.io.DataReader; @AllArgsConstructor public class NbtReader { private final DataReader reader; public byte readByte() throws IOException { return reader.readByte(); } public short readShort() throws IOException { return reader.readShort(); } public int readInt() throws IOException { return reader.readInt(); } public long readLong() throws IOException { return reader.readLong(); } public float readFloat() throws IOException { return reader.readFloat(); } public double readDouble() throws IOException { return reader.readDouble(); } public byte[] readByteArray() throws IOException { int length
= reader.readInt();
byte[] data = new byte[length]; reader.readBytes(data); return data; } public String readString() throws IOException { int length = reader.readShort(); byte[] data = new byte[length]; reader.readBytes(data); return new String(data, StandardCharsets.UTF_8); } @SuppressWarnings("unchecked") public NbtTagList<?> readList() throws IOException { byte type = reader.readByte(); int length = reader.readInt(); NbtTagList<Object> list = new NbtTagList<>(type, length); for (int i = 0; i < length; i++) { list.add((NbtTagValue<Object>) readTag(type)); } return list; } public NbtTagCompound readCompound() throws IOException { NbtTagCompound compound = new NbtTagCompound(); for (byte type; (type = reader.readByte()) != NbtTagValue.TAG_END;) { compound.put(readString(), readTag(type)); } return compound; } public @Nullable NbtRootCompound readRootCompoundOptional() throws IOException { byte type = reader.readByte(); if (type == NbtTagValue.TAG_END) return null; return readRootCompoundInner(type); } public NbtRootCompound readRootCompound() throws IOException { return readRootCompoundInner(reader.readByte()); } private NbtRootCompound readRootCompoundInner(byte type) throws IOException { if (type != NbtTagValue.TAG_COMPOUND) { throw new RuntimeException("Root is not TAG_Compound type"); } return new NbtRootCompound(readString(), readCompound()); } public int[] readIntArray() throws IOException { int length = reader.readInt(); int[] value = new int[length]; for (int i = 0; i < length; i++) { value[i] = reader.readInt(); } return value; } public long[] readLongArray() throws IOException { int length = reader.readInt(); long[] value = new long[length]; for (int i = 0; i < length; i++) { value[i] = reader.readLong(); } return value; } public NbtTagValue<?> readTag(byte type) throws IOException { switch (type) { case NbtTagValue.TAG_BYTE: return NbtTagValue.fromByte(readByte()); case NbtTagValue.TAG_SHORT: return NbtTagValue.fromShort(readShort()); case NbtTagValue.TAG_INT: return NbtTagValue.fromInt(readInt()); case NbtTagValue.TAG_LONG: return NbtTagValue.fromLong(readLong()); case NbtTagValue.TAG_FLOAT: return NbtTagValue.fromFloat(readFloat()); case NbtTagValue.TAG_DOUBLE: return NbtTagValue.fromDouble(readDouble()); case NbtTagValue.TAG_BYTE_ARRAY: return NbtTagValue.fromByteArray(readByteArray()); case NbtTagValue.TAG_STRING: return NbtTagValue.fromString(readString()); case NbtTagValue.TAG_LIST: return NbtTagValue.fromList(readList()); case NbtTagValue.TAG_COMPOUND: return NbtTagValue.fromCompound(readCompound()); case NbtTagValue.TAG_INT_ARRAY: return NbtTagValue.fromIntArray(readIntArray()); case NbtTagValue.TAG_LONG_ARRAY: return NbtTagValue.fromLongArray(readLongArray()); default: throw new RuntimeException("Unknown NBT type: " + type); } } }
api/src/main/java/sh/pancake/serdemc/data/nbt/io/NbtReader.java
storycraft-serde-mc-757a9e4
[ { "filename": "api/src/main/java/sh/pancake/serdemc/network/io/PacketDataReader.java", "retrieved_chunk": " }\n public short readShort() throws IOException {\n return reader.readShort();\n }\n public int readShortUnsigned() throws IOException {\n return reader.readShortUnsigned();\n }\n public int readInt() throws IOException {\n return reader.readInt();\n }", "score": 0.9280948042869568 }, { "filename": "api/src/main/java/sh/pancake/serdemc/io/DataReader.java", "retrieved_chunk": " default int readShortUnsigned() throws IOException {\n return Short.toUnsignedInt(readShort());\n }\n int readInt() throws IOException;\n long readLong() throws IOException;\n default float readFloat() throws IOException {\n return Float.intBitsToFloat(readInt());\n }\n default double readDouble() throws IOException {\n return Double.longBitsToDouble(readLong());", "score": 0.9140951037406921 }, { "filename": "api/src/main/java/sh/pancake/serdemc/io/InputStreamReader.java", "retrieved_chunk": " @Override\n public short readShort() throws IOException {\n return stream.readShort();\n }\n @Override\n public int readInt() throws IOException {\n return stream.readInt();\n }\n @Override\n public long readLong() throws IOException {", "score": 0.9112728238105774 }, { "filename": "api/src/main/java/sh/pancake/serdemc/network/io/PacketDataReader.java", "retrieved_chunk": " return reader.readByteUnsigned();\n }\n public boolean readBoolean() throws IOException {\n if (reader.readByte() == 0x00) {\n return false;\n }\n return true;\n }\n public int readBytes(byte[] dst) throws IOException {\n return reader.readBytes(dst);", "score": 0.9053330421447754 }, { "filename": "api/src/main/java/sh/pancake/serdemc/network/io/PacketDataReader.java", "retrieved_chunk": " public double readDouble() throws IOException {\n return reader.readDouble();\n }\n public String readString() throws IOException {\n return readString(32767);\n }\n public String readString(int maxLength) throws IOException {\n int length = readVarInt();\n if (length > maxLength) {\n throw new RuntimeException(\"String exceed max length. maxLength: \" + maxLength + \" length: \" + length);", "score": 0.9043182730674744 } ]
java
= reader.readInt();
/* * Created on Sun Mar 26 2023 * * Copyright (c) storycraft. Licensed under the Apache Licence 2.0. */ package sh.pancake.serdemc.data.nbt.io; import java.io.IOException; import java.nio.charset.StandardCharsets; import javax.annotation.Nullable; import lombok.AllArgsConstructor; import sh.pancake.serdemc.data.nbt.NbtRootCompound; import sh.pancake.serdemc.data.nbt.NbtTagCompound; import sh.pancake.serdemc.data.nbt.NbtTagList; import sh.pancake.serdemc.data.nbt.NbtTagValue; import sh.pancake.serdemc.io.DataReader; @AllArgsConstructor public class NbtReader { private final DataReader reader; public byte readByte() throws IOException { return reader.readByte(); } public short readShort() throws IOException { return reader.readShort(); } public int readInt() throws IOException { return reader.readInt(); } public long readLong() throws IOException { return reader.readLong(); } public float readFloat() throws IOException { return reader.readFloat(); } public double readDouble() throws IOException { return reader.readDouble(); } public byte[] readByteArray() throws IOException { int length = reader.readInt(); byte[] data = new byte[length];
reader.readBytes(data);
return data; } public String readString() throws IOException { int length = reader.readShort(); byte[] data = new byte[length]; reader.readBytes(data); return new String(data, StandardCharsets.UTF_8); } @SuppressWarnings("unchecked") public NbtTagList<?> readList() throws IOException { byte type = reader.readByte(); int length = reader.readInt(); NbtTagList<Object> list = new NbtTagList<>(type, length); for (int i = 0; i < length; i++) { list.add((NbtTagValue<Object>) readTag(type)); } return list; } public NbtTagCompound readCompound() throws IOException { NbtTagCompound compound = new NbtTagCompound(); for (byte type; (type = reader.readByte()) != NbtTagValue.TAG_END;) { compound.put(readString(), readTag(type)); } return compound; } public @Nullable NbtRootCompound readRootCompoundOptional() throws IOException { byte type = reader.readByte(); if (type == NbtTagValue.TAG_END) return null; return readRootCompoundInner(type); } public NbtRootCompound readRootCompound() throws IOException { return readRootCompoundInner(reader.readByte()); } private NbtRootCompound readRootCompoundInner(byte type) throws IOException { if (type != NbtTagValue.TAG_COMPOUND) { throw new RuntimeException("Root is not TAG_Compound type"); } return new NbtRootCompound(readString(), readCompound()); } public int[] readIntArray() throws IOException { int length = reader.readInt(); int[] value = new int[length]; for (int i = 0; i < length; i++) { value[i] = reader.readInt(); } return value; } public long[] readLongArray() throws IOException { int length = reader.readInt(); long[] value = new long[length]; for (int i = 0; i < length; i++) { value[i] = reader.readLong(); } return value; } public NbtTagValue<?> readTag(byte type) throws IOException { switch (type) { case NbtTagValue.TAG_BYTE: return NbtTagValue.fromByte(readByte()); case NbtTagValue.TAG_SHORT: return NbtTagValue.fromShort(readShort()); case NbtTagValue.TAG_INT: return NbtTagValue.fromInt(readInt()); case NbtTagValue.TAG_LONG: return NbtTagValue.fromLong(readLong()); case NbtTagValue.TAG_FLOAT: return NbtTagValue.fromFloat(readFloat()); case NbtTagValue.TAG_DOUBLE: return NbtTagValue.fromDouble(readDouble()); case NbtTagValue.TAG_BYTE_ARRAY: return NbtTagValue.fromByteArray(readByteArray()); case NbtTagValue.TAG_STRING: return NbtTagValue.fromString(readString()); case NbtTagValue.TAG_LIST: return NbtTagValue.fromList(readList()); case NbtTagValue.TAG_COMPOUND: return NbtTagValue.fromCompound(readCompound()); case NbtTagValue.TAG_INT_ARRAY: return NbtTagValue.fromIntArray(readIntArray()); case NbtTagValue.TAG_LONG_ARRAY: return NbtTagValue.fromLongArray(readLongArray()); default: throw new RuntimeException("Unknown NBT type: " + type); } } }
api/src/main/java/sh/pancake/serdemc/data/nbt/io/NbtReader.java
storycraft-serde-mc-757a9e4
[ { "filename": "api/src/main/java/sh/pancake/serdemc/network/io/PacketDataReader.java", "retrieved_chunk": " }\n public short readShort() throws IOException {\n return reader.readShort();\n }\n public int readShortUnsigned() throws IOException {\n return reader.readShortUnsigned();\n }\n public int readInt() throws IOException {\n return reader.readInt();\n }", "score": 0.9175057411193848 }, { "filename": "api/src/main/java/sh/pancake/serdemc/io/DataReader.java", "retrieved_chunk": " default int readShortUnsigned() throws IOException {\n return Short.toUnsignedInt(readShort());\n }\n int readInt() throws IOException;\n long readLong() throws IOException;\n default float readFloat() throws IOException {\n return Float.intBitsToFloat(readInt());\n }\n default double readDouble() throws IOException {\n return Double.longBitsToDouble(readLong());", "score": 0.9136583805084229 }, { "filename": "api/src/main/java/sh/pancake/serdemc/network/io/PacketDataReader.java", "retrieved_chunk": " public double readDouble() throws IOException {\n return reader.readDouble();\n }\n public String readString() throws IOException {\n return readString(32767);\n }\n public String readString(int maxLength) throws IOException {\n int length = readVarInt();\n if (length > maxLength) {\n throw new RuntimeException(\"String exceed max length. maxLength: \" + maxLength + \" length: \" + length);", "score": 0.9088525772094727 }, { "filename": "api/src/main/java/sh/pancake/serdemc/io/InputStreamReader.java", "retrieved_chunk": " @Override\n public short readShort() throws IOException {\n return stream.readShort();\n }\n @Override\n public int readInt() throws IOException {\n return stream.readInt();\n }\n @Override\n public long readLong() throws IOException {", "score": 0.9054248332977295 }, { "filename": "api/src/main/java/sh/pancake/serdemc/network/io/PacketDataReader.java", "retrieved_chunk": " return reader.readByteUnsigned();\n }\n public boolean readBoolean() throws IOException {\n if (reader.readByte() == 0x00) {\n return false;\n }\n return true;\n }\n public int readBytes(byte[] dst) throws IOException {\n return reader.readBytes(dst);", "score": 0.8983098268508911 } ]
java
reader.readBytes(data);
/* * Created on Sun Mar 26 2023 * * Copyright (c) storycraft. Licensed under the Apache Licence 2.0. */ package sh.pancake.serdemc.data.nbt.io; import java.io.IOException; import java.nio.charset.StandardCharsets; import javax.annotation.Nullable; import lombok.AllArgsConstructor; import sh.pancake.serdemc.data.nbt.NbtRootCompound; import sh.pancake.serdemc.data.nbt.NbtTagCompound; import sh.pancake.serdemc.data.nbt.NbtTagList; import sh.pancake.serdemc.data.nbt.NbtTagValue; import sh.pancake.serdemc.io.DataReader; @AllArgsConstructor public class NbtReader { private final DataReader reader; public byte readByte() throws IOException { return reader.readByte(); } public short readShort() throws IOException { return reader.readShort(); } public int readInt() throws IOException { return reader.readInt(); } public long readLong() throws IOException { return reader.readLong(); } public float readFloat() throws IOException { return reader.readFloat(); } public double readDouble() throws IOException { return reader.readDouble(); } public byte[] readByteArray() throws IOException { int length = reader.readInt(); byte[] data = new byte[length]; reader.readBytes(data); return data; } public String readString() throws IOException {
int length = reader.readShort();
byte[] data = new byte[length]; reader.readBytes(data); return new String(data, StandardCharsets.UTF_8); } @SuppressWarnings("unchecked") public NbtTagList<?> readList() throws IOException { byte type = reader.readByte(); int length = reader.readInt(); NbtTagList<Object> list = new NbtTagList<>(type, length); for (int i = 0; i < length; i++) { list.add((NbtTagValue<Object>) readTag(type)); } return list; } public NbtTagCompound readCompound() throws IOException { NbtTagCompound compound = new NbtTagCompound(); for (byte type; (type = reader.readByte()) != NbtTagValue.TAG_END;) { compound.put(readString(), readTag(type)); } return compound; } public @Nullable NbtRootCompound readRootCompoundOptional() throws IOException { byte type = reader.readByte(); if (type == NbtTagValue.TAG_END) return null; return readRootCompoundInner(type); } public NbtRootCompound readRootCompound() throws IOException { return readRootCompoundInner(reader.readByte()); } private NbtRootCompound readRootCompoundInner(byte type) throws IOException { if (type != NbtTagValue.TAG_COMPOUND) { throw new RuntimeException("Root is not TAG_Compound type"); } return new NbtRootCompound(readString(), readCompound()); } public int[] readIntArray() throws IOException { int length = reader.readInt(); int[] value = new int[length]; for (int i = 0; i < length; i++) { value[i] = reader.readInt(); } return value; } public long[] readLongArray() throws IOException { int length = reader.readInt(); long[] value = new long[length]; for (int i = 0; i < length; i++) { value[i] = reader.readLong(); } return value; } public NbtTagValue<?> readTag(byte type) throws IOException { switch (type) { case NbtTagValue.TAG_BYTE: return NbtTagValue.fromByte(readByte()); case NbtTagValue.TAG_SHORT: return NbtTagValue.fromShort(readShort()); case NbtTagValue.TAG_INT: return NbtTagValue.fromInt(readInt()); case NbtTagValue.TAG_LONG: return NbtTagValue.fromLong(readLong()); case NbtTagValue.TAG_FLOAT: return NbtTagValue.fromFloat(readFloat()); case NbtTagValue.TAG_DOUBLE: return NbtTagValue.fromDouble(readDouble()); case NbtTagValue.TAG_BYTE_ARRAY: return NbtTagValue.fromByteArray(readByteArray()); case NbtTagValue.TAG_STRING: return NbtTagValue.fromString(readString()); case NbtTagValue.TAG_LIST: return NbtTagValue.fromList(readList()); case NbtTagValue.TAG_COMPOUND: return NbtTagValue.fromCompound(readCompound()); case NbtTagValue.TAG_INT_ARRAY: return NbtTagValue.fromIntArray(readIntArray()); case NbtTagValue.TAG_LONG_ARRAY: return NbtTagValue.fromLongArray(readLongArray()); default: throw new RuntimeException("Unknown NBT type: " + type); } } }
api/src/main/java/sh/pancake/serdemc/data/nbt/io/NbtReader.java
storycraft-serde-mc-757a9e4
[ { "filename": "api/src/main/java/sh/pancake/serdemc/network/io/PacketDataReader.java", "retrieved_chunk": " }\n public short readShort() throws IOException {\n return reader.readShort();\n }\n public int readShortUnsigned() throws IOException {\n return reader.readShortUnsigned();\n }\n public int readInt() throws IOException {\n return reader.readInt();\n }", "score": 0.9244194030761719 }, { "filename": "api/src/main/java/sh/pancake/serdemc/network/io/PacketDataReader.java", "retrieved_chunk": " public double readDouble() throws IOException {\n return reader.readDouble();\n }\n public String readString() throws IOException {\n return readString(32767);\n }\n public String readString(int maxLength) throws IOException {\n int length = readVarInt();\n if (length > maxLength) {\n throw new RuntimeException(\"String exceed max length. maxLength: \" + maxLength + \" length: \" + length);", "score": 0.9185706377029419 }, { "filename": "api/src/main/java/sh/pancake/serdemc/network/io/PacketDataReader.java", "retrieved_chunk": " return reader.readByteUnsigned();\n }\n public boolean readBoolean() throws IOException {\n if (reader.readByte() == 0x00) {\n return false;\n }\n return true;\n }\n public int readBytes(byte[] dst) throws IOException {\n return reader.readBytes(dst);", "score": 0.9172124862670898 }, { "filename": "api/src/main/java/sh/pancake/serdemc/io/InputStreamReader.java", "retrieved_chunk": " @Override\n public short readShort() throws IOException {\n return stream.readShort();\n }\n @Override\n public int readInt() throws IOException {\n return stream.readInt();\n }\n @Override\n public long readLong() throws IOException {", "score": 0.9007211327552795 }, { "filename": "api/src/main/java/sh/pancake/serdemc/io/DataReader.java", "retrieved_chunk": " return Byte.toUnsignedInt(readByte());\n }\n int readBytes(byte[] dst, int offset, int length) throws IOException;\n default int readBytes(byte[] dst, int offset) throws IOException {\n return readBytes(dst, offset, dst.length);\n }\n default int readBytes(byte[] dst) throws IOException {\n return readBytes(dst, 0, dst.length);\n }\n short readShort() throws IOException;", "score": 0.8999140858650208 } ]
java
int length = reader.readShort();
package com.solodroid.ads.sdkdemo.activity; import static com.solodroid.ads.sdk.util.Constant.ADMOB; import static com.solodroid.ads.sdk.util.Constant.AD_STATUS_ON; import static com.solodroid.ads.sdk.util.Constant.APPLOVIN; import static com.solodroid.ads.sdk.util.Constant.APPLOVIN_MAX; import static com.solodroid.ads.sdk.util.Constant.GOOGLE_AD_MANAGER; import static com.solodroid.ads.sdk.util.Constant.WORTISE; import android.app.Application; import android.content.Intent; import android.os.Bundle; import android.os.Handler; import android.os.Looper; import android.util.Log; import androidx.annotation.NonNull; import androidx.appcompat.app.AppCompatActivity; import com.solodroid.ads.sdk.format.AdNetwork; import com.solodroid.ads.sdk.format.AppOpenAd; import com.solodroid.ads.sdkdemo.BuildConfig; import com.solodroid.ads.sdkdemo.R; import com.solodroid.ads.sdkdemo.application.MyApplication; import com.solodroid.ads.sdkdemo.callback.CallbackConfig; import com.solodroid.ads.sdkdemo.data.Constant; import com.solodroid.ads.sdkdemo.database.SharedPref; import com.solodroid.ads.sdkdemo.rest.RestAdapter; import java.util.Arrays; import java.util.List; import retrofit2.Call; import retrofit2.Callback; import retrofit2.Response; @SuppressWarnings("ConstantConditions") public class ActivitySplash extends AppCompatActivity { private static final String TAG = "ActivitySplash"; Call<CallbackConfig> callbackConfigCall = null; public static int DELAY_PROGRESS = 1500; AdNetwork.Initialize adNetwork; AppOpenAd.Builder appOpenAdBuilder; SharedPref sharedPref; @Override protected void onCreate(Bundle savedInstanceState) { super.onCreate(savedInstanceState); setContentView(R.layout.activity_splash); sharedPref = new SharedPref(this); initAds(); if (Constant.AD_STATUS.equals(AD_STATUS_ON) && Constant.OPEN_ADS_ON_START) { if (!Constant.FORCE_TO_SHOW_APP_OPEN_AD_ON_START) { new Handler(Looper.getMainLooper()).postDelayed(() -> { switch (Constant.AD_NETWORK) { case ADMOB: if (!Constant.ADMOB_APP_OPEN_AD_ID.equals("0")) { ((MyApplication) getApplication()).showAdIfAvailable(ActivitySplash.this, this::requestConfig); } else { requestConfig(); } break; case GOOGLE_AD_MANAGER: if (!Constant.GOOGLE_AD_MANAGER_APP_OPEN_AD_ID.equals("0")) { ((MyApplication) getApplication()).showAdIfAvailable(ActivitySplash.this, this::requestConfig); } else { requestConfig(); } break; case APPLOVIN: case APPLOVIN_MAX: if (!Constant.APPLOVIN_APP_OPEN_AP_ID.equals("0")) { ((MyApplication) getApplication()).showAdIfAvailable(ActivitySplash.this, this::requestConfig); } else { requestConfig(); } break; case WORTISE: if (!Constant.WORTISE_APP_OPEN_AD_ID.equals("0")) { ((MyApplication) getApplication()).showAdIfAvailable(ActivitySplash.this, this::requestConfig); } else { requestConfig(); } break; default: requestConfig(); break; } }, DELAY_PROGRESS); } else { requestConfig(); } } else { requestConfig(); } } private void requestConfig() { requestAPI("https://raw.githubusercontent.com/solodroidev/content/uploads/json/android.json"); } private void requestAPI(@SuppressWarnings("SameParameterValue") String url) { if (url.startsWith("http://") || url.startsWith("https://")) { if (url.contains("https://drive.google.com")) { String driveUrl = url.replace("https://", "").replace("http://", ""); List<String> data = Arrays.asList(driveUrl.split("/")); String googleDriveFileId = data.get(3); callbackConfigCall
= RestAdapter.createApi().getDriveJsonFileId(googleDriveFileId);
} else { callbackConfigCall = RestAdapter.createApi().getJsonUrl(url); } } else { callbackConfigCall = RestAdapter.createApi().getDriveJsonFileId(url); } callbackConfigCall.enqueue(new Callback<CallbackConfig>() { public void onResponse(@NonNull Call<CallbackConfig> call, @NonNull Response<CallbackConfig> response) { CallbackConfig resp = response.body(); if (resp != null) { sharedPref.savePostList(resp.android); loadOpenAds(); Log.d(TAG, "responses success"); } else { loadOpenAds(); Log.d(TAG, "responses null"); } } public void onFailure(@NonNull Call<CallbackConfig> call, @NonNull Throwable th) { Log.d(TAG, "responses failed: " + th.getMessage()); loadOpenAds(); } }); } private void initAds() { adNetwork = new AdNetwork.Initialize(this) .setAdStatus(Constant.AD_STATUS) .setAdNetwork(Constant.AD_NETWORK) .setBackupAdNetwork(Constant.BACKUP_AD_NETWORK) .setAdMobAppId(null) .setStartappAppId(Constant.STARTAPP_APP_ID) .setUnityGameId(Constant.UNITY_GAME_ID) .setAppLovinSdkKey(getResources().getString(R.string.applovin_sdk_key)) .setIronSourceAppKey(Constant.IRONSOURCE_APP_KEY) .setWortiseAppId(Constant.WORTISE_APP_ID) .setDebug(BuildConfig.DEBUG) .build(); } private void loadOpenAds() { if (Constant.FORCE_TO_SHOW_APP_OPEN_AD_ON_START && Constant.OPEN_ADS_ON_START) { appOpenAdBuilder = new AppOpenAd.Builder(this) .setAdStatus(Constant.AD_STATUS) .setAdNetwork(Constant.AD_NETWORK) .setBackupAdNetwork(Constant.BACKUP_AD_NETWORK) .setAdMobAppOpenId(Constant.ADMOB_APP_OPEN_AD_ID) .setAdManagerAppOpenId(Constant.GOOGLE_AD_MANAGER_APP_OPEN_AD_ID) .setApplovinAppOpenId(Constant.APPLOVIN_APP_OPEN_AP_ID) .setWortiseAppOpenId(Constant.WORTISE_APP_OPEN_AD_ID) .build(this::startMainActivity); } else { startMainActivity(); } } public void startMainActivity() { new Handler().postDelayed(() -> { Intent intent = new Intent(this, MainActivity.class); startActivity(intent); finish(); }, DELAY_PROGRESS); } }
demo/src/main/java/com/solodroid/ads/sdkdemo/activity/ActivitySplash.java
solodroidx-multi-ads-sdk-4f23cdf
[ { "filename": "demo/src/main/java/com/solodroid/ads/sdkdemo/rest/ApiInterface.java", "retrieved_chunk": " @Headers({CACHE, AGENT})\n @GET\n Call<CallbackConfig> getJsonUrl(\n @Url String url\n );\n @Headers({CACHE, AGENT})\n @GET(\"uc?export=download\")\n Call<CallbackConfig> getDriveJsonFileId(\n @Query(\"id\") String id\n );", "score": 0.8838316798210144 }, { "filename": "demo/src/main/java/com/solodroid/ads/sdkdemo/rest/RestAdapter.java", "retrieved_chunk": " OkHttpClient okHttpClient = new OkHttpClient.Builder()\n .connectTimeout(5, TimeUnit.SECONDS)\n .writeTimeout(10, TimeUnit.SECONDS)\n .readTimeout(30, TimeUnit.SECONDS)\n .cache(null)\n .build();\n Retrofit retrofit = new Retrofit.Builder()\n .baseUrl(\"https://drive.google.com/\")\n .addConverterFactory(GsonConverterFactory.create())\n .client(okHttpClient)", "score": 0.8388479948043823 }, { "filename": "demo/src/main/java/com/solodroid/ads/sdkdemo/rest/ApiInterface.java", "retrieved_chunk": "package com.solodroid.ads.sdkdemo.rest;\nimport com.solodroid.ads.sdkdemo.callback.CallbackConfig;\nimport retrofit2.Call;\nimport retrofit2.http.GET;\nimport retrofit2.http.Headers;\nimport retrofit2.http.Query;\nimport retrofit2.http.Url;\npublic interface ApiInterface {\n String CACHE = \"Cache-Control: max-age=0\";\n String AGENT = \"Data-Agent: Solodroid\";", "score": 0.8089445233345032 }, { "filename": "demo/src/main/java/com/solodroid/ads/sdkdemo/rest/RestAdapter.java", "retrieved_chunk": "package com.solodroid.ads.sdkdemo.rest;\nimport java.util.concurrent.TimeUnit;\nimport okhttp3.OkHttpClient;\nimport okhttp3.logging.HttpLoggingInterceptor;\nimport retrofit2.Retrofit;\nimport retrofit2.converter.gson.GsonConverterFactory;\npublic class RestAdapter {\n public static ApiInterface createApi() {\n HttpLoggingInterceptor logging = new HttpLoggingInterceptor();\n logging.setLevel(HttpLoggingInterceptor.Level.BODY);", "score": 0.796228289604187 }, { "filename": "library/src/main/java/com/solodroid/ads/sdk/format/AppOpenAd.java", "retrieved_chunk": " });\n break;\n case GOOGLE_AD_MANAGER:\n case FAN_BIDDING_AD_MANAGER:\n @SuppressLint(\"VisibleForTests\") AdManagerAdRequest adManagerAdRequest = new AdManagerAdRequest.Builder().build();\n com.google.android.gms.ads.appopen.AppOpenAd.load(activity, adManagerAppOpenId, adManagerAdRequest, new com.google.android.gms.ads.appopen.AppOpenAd.AppOpenAdLoadCallback() {\n @Override\n public void onAdLoaded(@NonNull com.google.android.gms.ads.appopen.AppOpenAd ad) {\n appOpenAd = ad;\n isAppOpenAdLoaded = true;", "score": 0.7812830209732056 } ]
java
= RestAdapter.createApi().getDriveJsonFileId(googleDriveFileId);
/* * Copyright Strimzi authors. * License: Apache License 2.0 (see the file LICENSE or http://apache.org/licenses/LICENSE-2.0.html). */ package io.strimzi.kafka.bridge.mqtt.mapper; import java.util.HashMap; import java.util.List; import java.util.Map; import java.util.regex.Matcher; import java.util.regex.Pattern; /** * Responsible for handling all the topic mapping using named placeholders instead of regular expressions. * * @see MappingRule * @see MqttKafkaMapper * @see MqttKafkaRegexMapper */ public class MqttKafkaSimpleMapper extends MqttKafkaMapper { // find any word inside a curly bracket. E.g. {something}, this is known as a placeholder. private static final String MQTT_TOPIC_PLACEHOLDER_REGEX = "\\{\\w+\\}"; // identifies a single level wildcard character in the mqtt pattern. E.g. sensors/+/data private static final String MQTT_TOPIC_SINGLE_LEVEL_WILDCARD_CHARACTER = "+"; // Regex expression used to replace the + in the mqtt pattern. private static final String SINGLE_LEVEL_WILDCARD_REGEX = "[^/]+"; // identifies a multi level wildcard character in the mqtt pattern. E.g. sensors/# private static final String MQTT_TOPIC_MULTI_LEVEL_WILDCARD_CHARACTER = "#"; // used to replace the # in the mqtt pattern. public static final String WILDCARD_REGEX = "(?:\\/.*)?$"; /** * Constructor. * * @param rules the list of mapping rules. */ public MqttKafkaSimpleMapper(List<MappingRule> rules) { super(rules, Pattern.compile(MQTT_TOPIC_PLACEHOLDER_REGEX)); } @Override public MappingResult map(String mqttTopic) { for (MappingRule rule : this.rules) { Matcher matcher = this.patterns.get(this.rules.indexOf(rule)).matcher(mqttTopic); if (matcher.matches()) { HashMap<String, String> placeholders = new HashMap<>(); String mappedKafkaTopic = rule.getKafkaTopicTemplate(); String kafkaKey = rule.getKafkaKeyTemplate(); // find MQTT_TOPIC_PLACEHOLDER_REGEX in the kafkaTopicTemplate. Matcher placeholderMatcher = this.placeholderPattern
.matcher(rule.getKafkaTopicTemplate());
while (placeholderMatcher.find()) { String placeholderKey = placeholderMatcher.group(); placeholders.put(placeholderKey, null); } // find MQTT_TOPIC_PLACEHOLDER_REGEX in the kafkaKey if (kafkaKey != null) { placeholderMatcher = this.placeholderPattern.matcher(kafkaKey); while (placeholderMatcher.find()) { String placeholderKey = placeholderMatcher.group(); placeholders.put(placeholderKey, null); } } if (!placeholders.isEmpty()) { Matcher mqttTopicMatcher = this.placeholderPattern.matcher(rule.getMqttTopicPattern()); // find the placeholders in the mqtt topic pattern and assign them a value. while (mqttTopicMatcher.find()) { String placeholderKey = mqttTopicMatcher.group(); String placeholderValue = matcher.group(removeBrackets(placeholderKey)); placeholders.put(placeholderKey, placeholderValue); } // build the Kafka topic using the placeholders. for (Map.Entry<String, String> entry : placeholders.entrySet()) { if (entry.getValue() != null) { mappedKafkaTopic = mappedKafkaTopic.replace(entry.getKey(), entry.getValue()); kafkaKey = kafkaKey != null ? kafkaKey.replace(entry.getKey(), entry.getValue()) : null; } else { throw new IllegalArgumentException("The placeholder " + entry.getKey() + " was not found assigned any value."); } } } return new MappingResult(mappedKafkaTopic, kafkaKey); } } return new MappingResult(MqttKafkaMapper.DEFAULT_KAFKA_TOPIC, null); } @Override protected void buildOrCompilePatterns() { // convert the mqtt patterns to a valid regex expression. // the mqtt pattern can contain placeholders like {something}, + and #. // if the mqtt topic contains a +, we replace it with @singleLevelWildcardRegex // if the mqtt topic contains a #, we replace it with @multiLevelWildcardRegex // if the mqtt topic contains a placeholder (pattern \{\w+\}), we replace it with @placeholderRegex String[] mqttTopicPatternParts; StringBuilder ruleRegex; for (MappingRule rule : this.rules) { mqttTopicPatternParts = rule.getMqttTopicPattern().split(MQTT_TOPIC_SEPARATOR); ruleRegex = new StringBuilder(); for (String part : mqttTopicPatternParts) { if (part.matches(MQTT_TOPIC_PLACEHOLDER_REGEX)) { ruleRegex.append(buildNamedRegexExpression(part)); } else if (part.equals(MQTT_TOPIC_SINGLE_LEVEL_WILDCARD_CHARACTER)) { ruleRegex.append(SINGLE_LEVEL_WILDCARD_REGEX); } else if (part.equals(MQTT_TOPIC_MULTI_LEVEL_WILDCARD_CHARACTER)) { if (ruleRegex.length() > 1) { ruleRegex.deleteCharAt(ruleRegex.length() - 1); } ruleRegex.append(WILDCARD_REGEX); } else { ruleRegex.append(part); } ruleRegex.append(MQTT_TOPIC_SEPARATOR); } // remove the last slash ruleRegex.deleteCharAt(ruleRegex.length() - 1); // compile the regex expression for the rule. patterns.add(Pattern.compile(ruleRegex.toString())); } } /** * Helper method for building a named regex expression. * A named regex expression is a regex expression that contains a named capturing group. * E.g. (?<groupName>regexExpression) * * @param placeholder represents a placeholder in the mqtt pattern. * @return a named regex expression. */ private String buildNamedRegexExpression(String placeholder) { String groupName = removeBrackets(placeholder); return "(?<" + groupName + ">[^/]+)"; } /** * Helper method for removing the curly brackets from a placeholder. * * @param placeholder represents a placeholder in the pattern. * @return a placeholder without the curly brackets. */ private String removeBrackets(String placeholder) { return placeholder.replaceAll("\\{+|\\}+", ""); } }
src/main/java/io/strimzi/kafka/bridge/mqtt/mapper/MqttKafkaSimpleMapper.java
strimzi-strimzi-mqtt-bridge-cb0a4b8
[ { "filename": "src/main/java/io/strimzi/kafka/bridge/mqtt/mapper/MqttKafkaRegexMapper.java", "retrieved_chunk": " }\n @Override\n public MappingResult map(String mqttTopic) {\n for (MappingRule rule : this.rules) {\n Matcher matcher = this.patterns.get(this.rules.indexOf(rule)).matcher(mqttTopic);\n if (matcher.matches()) {\n String mappedKafkaTopic = rule.getKafkaTopicTemplate();\n String kafkaKey = rule.getKafkaKeyTemplate();\n for (int i = 1; i < matcher.groupCount() + 1; i++) {\n mappedKafkaTopic = mappedKafkaTopic.replace(\"$\" + i, matcher.group(i));", "score": 0.9660855531692505 }, { "filename": "src/main/java/io/strimzi/kafka/bridge/mqtt/mapper/MqttKafkaMapper.java", "retrieved_chunk": " }\n /**\n * Maps an MQTT topic to a Kafka topic. The topic is mapped according to the defined mapping rules.\n *\n * @param mqttTopic\n * @return a MappingResult object containing the mapped Kafka topic and Kafka key.\n */\n public abstract MappingResult map(String mqttTopic);\n /**\n * Helper method for Building the regex expressions for the mapping rules.", "score": 0.8989261984825134 }, { "filename": "src/main/java/io/strimzi/kafka/bridge/mqtt/mapper/MqttKafkaRegexMapper.java", "retrieved_chunk": " }\n }\n return new MappingResult(MqttKafkaMapper.DEFAULT_KAFKA_TOPIC, null);\n }\n @Override\n protected void buildOrCompilePatterns() {\n this.rules.forEach(rule-> this.patterns.add(Pattern.compile(rule.getMqttTopicPattern())));\n }\n /**\n * Checks if there are any pending placeholders in the Kafka topic or Kafka key template.", "score": 0.8801529407501221 }, { "filename": "src/main/java/io/strimzi/kafka/bridge/mqtt/mapper/MqttKafkaMapper.java", "retrieved_chunk": " */\npublic abstract class MqttKafkaMapper {\n // default Kafka topic. Used when no mapping rule matches the mqtt topic.\n public static final String DEFAULT_KAFKA_TOPIC = \"messages_default\";\n // MQTT topic separator\n public static final String MQTT_TOPIC_SEPARATOR = \"/\";\n protected final List<MappingRule> rules;\n protected final List<Pattern> patterns = new ArrayList<>();\n protected final Pattern placeholderPattern;\n /**", "score": 0.8798171877861023 }, { "filename": "src/test/java/io/strimzi/kafka/bridge/mqtt/mapper/MqttKafkaRegexMapperTest.java", "retrieved_chunk": " rules.add(new MappingRule(\"([^/]+)/.*\", \"$1\", null));\n MqttKafkaRegexMapper mapper = new MqttKafkaRegexMapper(rules);\n // Test for building/([^/]+)/room/(\\\\d{1,3}).* pattern\n MappingResult mappingResult = mapper.map(\"building/4/room/23/temperature\");\n assertThat(\"Mqtt pattern building/([^/]+)/room/(\\\\d{1,3}).* should be mapped to building_$1\",\n mappingResult.kafkaTopic(), is(\"building_4\"));\n assertThat(\"The key for building/([^/]+)/room/(\\\\d{1,3}).* should be expanded to room_$2\",\n mappingResult.kafkaKey(), is(\"room_23\"));\n // Test for building/([^/]+).* pattern\n mappingResult = mapper.map(\"building/405/room\");", "score": 0.8764773011207581 } ]
java
.matcher(rule.getKafkaTopicTemplate());
/* * Copyright Strimzi authors. * License: Apache License 2.0 (see the file LICENSE or http://apache.org/licenses/LICENSE-2.0.html). */ package io.strimzi.kafka.bridge.mqtt.kafka; import io.strimzi.kafka.bridge.mqtt.config.KafkaConfig; import org.apache.kafka.clients.producer.KafkaProducer; import org.apache.kafka.clients.producer.ProducerConfig; import org.apache.kafka.clients.producer.ProducerRecord; import org.apache.kafka.clients.producer.RecordMetadata; import org.apache.kafka.clients.producer.Producer; import org.apache.kafka.common.serialization.ByteArraySerializer; import org.apache.kafka.common.serialization.StringSerializer; import java.util.Properties; import java.util.concurrent.CompletableFuture; import java.util.concurrent.CompletionStage; /** * Represents a Kafka producer for the Bridge. */ public class KafkaBridgeProducer { private final Producer<String, byte[]> noAckProducer; private final Producer<String, byte[]> ackOneProducer; /** * Constructor */ public KafkaBridgeProducer(KafkaConfig config) { this.noAckProducer = createProducer(config, KafkaProducerAckLevel.ZERO); this.ackOneProducer = createProducer(config, KafkaProducerAckLevel.ONE); } /** * Send the given record to the Kafka topic * * @param record record to be sent * @return a future which completes when the record is acknowledged */ public CompletionStage<RecordMetadata> send(ProducerRecord<String, byte[]> record) { CompletableFuture<RecordMetadata> promise = new CompletableFuture<>(); this.ackOneProducer.send(record, (metadata, exception) -> { if (exception != null) { promise.completeExceptionally(exception); } else { promise.complete(metadata); } }); return promise; } /** * Send the given record to the Kafka topic * * @param record record to be sent */ public void sendNoAck(ProducerRecord<String, byte[]> record) { this.noAckProducer.send(record); } /** * Create the Kafka producer client with the given configuration */ private Producer<String, byte[]> createProducer(KafkaConfig kafkaConfig, KafkaProducerAckLevel producerAckLevel) { Properties props = new Properties(); props.putAll(kafkaConfig.getConfig()); props.putAll(
kafkaConfig.getProducerConfig().getConfig());
props.put(ProducerConfig.ACKS_CONFIG, String.valueOf(producerAckLevel.getValue())); return new KafkaProducer<>(props, new StringSerializer(), new ByteArraySerializer()); } /** * Close the producer */ public void close() { if (this.noAckProducer != null) { this.noAckProducer.flush(); this.noAckProducer.close(); } if (this.ackOneProducer != null) { this.ackOneProducer.flush(); this.ackOneProducer.close(); } } }
src/main/java/io/strimzi/kafka/bridge/mqtt/kafka/KafkaBridgeProducer.java
strimzi-strimzi-mqtt-bridge-cb0a4b8
[ { "filename": "src/test/java/io/strimzi/kafka/bridge/mqtt/kafka/KafkaBridgeProducerTest.java", "retrieved_chunk": " * Test the {@link KafkaBridgeProducer#send(ProducerRecord)}} method\n */\n @Test\n public void testSend() {\n // mock the producer\n KafkaBridgeProducer producer = mock(KafkaBridgeProducer.class);\n String kafkaTopic = \"test-topic\";\n ProducerRecord<String, byte[]> record = new ProducerRecord<>(kafkaTopic, \"test\".getBytes());\n // simulate the send method with ack\n when(producer.send(any(ProducerRecord.class)))", "score": 0.8374298810958862 }, { "filename": "src/test/java/io/strimzi/kafka/bridge/mqtt/MqttBridgetIT.java", "retrieved_chunk": " }\n // instantiate the kafka consumer client\n kafkaConsumerClient = new KafkaConsumer<>(\n ImmutableMap.of(\n ConsumerConfig.BOOTSTRAP_SERVERS_CONFIG, kafkaBootstrapServers,\n ConsumerConfig.GROUP_ID_CONFIG, \"gid-\" + UUID.randomUUID(),\n ConsumerConfig.AUTO_OFFSET_RESET_CONFIG, OffsetResetStrategy.EARLIEST.name().toLowerCase(Locale.ROOT)\n ),\n new StringDeserializer(),\n new StringDeserializer()", "score": 0.8369492888450623 }, { "filename": "src/test/java/io/strimzi/kafka/bridge/mqtt/MqttBridgetIT.java", "retrieved_chunk": " );\n // consumer client subscribes to the kafka topic\n kafkaConsumerClient.subscribe(Collections.singletonList(KAFKA_TOPIC));\n // prepare the configuration for the bridge\n BridgeConfig bridgeConfig = BridgeConfig.fromMap(\n ImmutableMap.of(\n BridgeConfig.BRIDGE_ID, \"my-bridge\",\n KafkaConfig.KAFKA_CONFIG_PREFIX + ConsumerConfig.BOOTSTRAP_SERVERS_CONFIG, kafkaBootstrapServers,\n MqttConfig.MQTT_HOST, MQTT_SERVER_HOST,\n MqttConfig.MQTT_PORT, MQTT_SERVER_PORT", "score": 0.822140634059906 }, { "filename": "src/test/java/io/strimzi/kafka/bridge/mqtt/MqttBridgetIT.java", "retrieved_chunk": " MqttConnectOptions options = new MqttConnectOptions();\n client.connect(options).waitForCompletion();\n client.publish(mqttTopic, \"Hello world\".getBytes(), MqttQoS.AT_MOST_ONCE.value(), false).waitForCompletion();\n ConsumerRecords<String, String> records = kafkaConsumerClient.poll(Duration.ofSeconds(20));\n assertThat(\"The record should be present\", records.count(), is(1));\n ConsumerRecord<String, String> record = records.iterator().next();\n assertThat(\"The key should be \" + kafkaKey,\n record.key(), is(kafkaKey));\n assertThat(\"The value should be Hello world\",\n record.value(), is(\"Hello world\"));", "score": 0.8217470645904541 }, { "filename": "src/main/java/io/strimzi/kafka/bridge/mqtt/config/KafkaProducerConfig.java", "retrieved_chunk": " */\npublic class KafkaProducerConfig extends AbstractConfig {\n // Prefix for all the specific configuration parameters for Kafka producer in the properties file\n public static final String KAFKA_PRODUCER_CONFIG_PREFIX = KafkaConfig.KAFKA_CONFIG_PREFIX + \"producer.\";\n /**\n * Constructor\n *\n * @param config configuration parameters map\n */\n public KafkaProducerConfig(Map<String, Object> config) {", "score": 0.8207722306251526 } ]
java
kafkaConfig.getProducerConfig().getConfig());
/* * Copyright Strimzi authors. * License: Apache License 2.0 (see the file LICENSE or http://apache.org/licenses/LICENSE-2.0.html). */ package io.strimzi.kafka.bridge.mqtt.kafka; import io.strimzi.kafka.bridge.mqtt.config.KafkaConfig; import org.apache.kafka.clients.producer.KafkaProducer; import org.apache.kafka.clients.producer.ProducerConfig; import org.apache.kafka.clients.producer.ProducerRecord; import org.apache.kafka.clients.producer.RecordMetadata; import org.apache.kafka.clients.producer.Producer; import org.apache.kafka.common.serialization.ByteArraySerializer; import org.apache.kafka.common.serialization.StringSerializer; import java.util.Properties; import java.util.concurrent.CompletableFuture; import java.util.concurrent.CompletionStage; /** * Represents a Kafka producer for the Bridge. */ public class KafkaBridgeProducer { private final Producer<String, byte[]> noAckProducer; private final Producer<String, byte[]> ackOneProducer; /** * Constructor */ public KafkaBridgeProducer(KafkaConfig config) { this.noAckProducer = createProducer(config, KafkaProducerAckLevel.ZERO); this.ackOneProducer = createProducer(config, KafkaProducerAckLevel.ONE); } /** * Send the given record to the Kafka topic * * @param record record to be sent * @return a future which completes when the record is acknowledged */ public CompletionStage<RecordMetadata> send(ProducerRecord<String, byte[]> record) { CompletableFuture<RecordMetadata> promise = new CompletableFuture<>(); this.ackOneProducer.send(record, (metadata, exception) -> { if (exception != null) { promise.completeExceptionally(exception); } else { promise.complete(metadata); } }); return promise; } /** * Send the given record to the Kafka topic * * @param record record to be sent */ public void sendNoAck(ProducerRecord<String, byte[]> record) { this.noAckProducer.send(record); } /** * Create the Kafka producer client with the given configuration */ private Producer<String, byte[]> createProducer(KafkaConfig kafkaConfig, KafkaProducerAckLevel producerAckLevel) { Properties props = new Properties(); props.putAll(kafkaConfig.getConfig()); props.putAll(kafkaConfig.getProducerConfig().getConfig()); props.put(ProducerConfig.
ACKS_CONFIG, String.valueOf(producerAckLevel.getValue()));
return new KafkaProducer<>(props, new StringSerializer(), new ByteArraySerializer()); } /** * Close the producer */ public void close() { if (this.noAckProducer != null) { this.noAckProducer.flush(); this.noAckProducer.close(); } if (this.ackOneProducer != null) { this.ackOneProducer.flush(); this.ackOneProducer.close(); } } }
src/main/java/io/strimzi/kafka/bridge/mqtt/kafka/KafkaBridgeProducer.java
strimzi-strimzi-mqtt-bridge-cb0a4b8
[ { "filename": "src/test/java/io/strimzi/kafka/bridge/mqtt/MqttBridgetIT.java", "retrieved_chunk": " }\n // instantiate the kafka consumer client\n kafkaConsumerClient = new KafkaConsumer<>(\n ImmutableMap.of(\n ConsumerConfig.BOOTSTRAP_SERVERS_CONFIG, kafkaBootstrapServers,\n ConsumerConfig.GROUP_ID_CONFIG, \"gid-\" + UUID.randomUUID(),\n ConsumerConfig.AUTO_OFFSET_RESET_CONFIG, OffsetResetStrategy.EARLIEST.name().toLowerCase(Locale.ROOT)\n ),\n new StringDeserializer(),\n new StringDeserializer()", "score": 0.8490355610847473 }, { "filename": "src/main/java/io/strimzi/kafka/bridge/mqtt/config/KafkaConfig.java", "retrieved_chunk": " *\n * @param config configuration parameters map\n * @param kafkaProducerConfig Kafka producer configuration properties\n */\n public KafkaConfig(Map<String, Object> config, KafkaProducerConfig kafkaProducerConfig) {\n super(config);\n this.kafkaProducerConfig = kafkaProducerConfig;\n }\n /**\n * Build a Kafka configuration object from a map of configuration parameters", "score": 0.8310028314590454 }, { "filename": "src/test/java/io/strimzi/kafka/bridge/mqtt/MqttBridgetIT.java", "retrieved_chunk": " );\n // consumer client subscribes to the kafka topic\n kafkaConsumerClient.subscribe(Collections.singletonList(KAFKA_TOPIC));\n // prepare the configuration for the bridge\n BridgeConfig bridgeConfig = BridgeConfig.fromMap(\n ImmutableMap.of(\n BridgeConfig.BRIDGE_ID, \"my-bridge\",\n KafkaConfig.KAFKA_CONFIG_PREFIX + ConsumerConfig.BOOTSTRAP_SERVERS_CONFIG, kafkaBootstrapServers,\n MqttConfig.MQTT_HOST, MQTT_SERVER_HOST,\n MqttConfig.MQTT_PORT, MQTT_SERVER_PORT", "score": 0.8271443843841553 }, { "filename": "src/main/java/io/strimzi/kafka/bridge/mqtt/config/KafkaProducerConfig.java", "retrieved_chunk": " */\npublic class KafkaProducerConfig extends AbstractConfig {\n // Prefix for all the specific configuration parameters for Kafka producer in the properties file\n public static final String KAFKA_PRODUCER_CONFIG_PREFIX = KafkaConfig.KAFKA_CONFIG_PREFIX + \"producer.\";\n /**\n * Constructor\n *\n * @param config configuration parameters map\n */\n public KafkaProducerConfig(Map<String, Object> config) {", "score": 0.821891725063324 }, { "filename": "src/main/java/io/strimzi/kafka/bridge/mqtt/config/KafkaConfig.java", "retrieved_chunk": " * @see AbstractConfig\n */\npublic class KafkaConfig extends AbstractConfig {\n // Prefix for all the specific configuration parameters for Kafka in the properties file\n public static final String KAFKA_CONFIG_PREFIX = \"kafka.\";\n // Kafka bootstrap server configuration. This will be removed when we add the apache kafka client\n public static final String BOOTSTRAP_SERVERS_CONFIG = KAFKA_CONFIG_PREFIX + \"bootstrap.servers\";\n private final KafkaProducerConfig kafkaProducerConfig;\n /**\n * Constructor", "score": 0.8175000548362732 } ]
java
ACKS_CONFIG, String.valueOf(producerAckLevel.getValue()));
/* * Created on Sun Mar 26 2023 * * Copyright (c) storycraft. Licensed under the Apache Licence 2.0. */ package sh.pancake.serdemc.data.nbt.io; import java.io.IOException; import java.nio.charset.StandardCharsets; import javax.annotation.Nullable; import lombok.AllArgsConstructor; import sh.pancake.serdemc.data.nbt.NbtRootCompound; import sh.pancake.serdemc.data.nbt.NbtTagCompound; import sh.pancake.serdemc.data.nbt.NbtTagList; import sh.pancake.serdemc.data.nbt.NbtTagValue; import sh.pancake.serdemc.io.DataReader; @AllArgsConstructor public class NbtReader { private final DataReader reader; public byte readByte() throws IOException { return reader.readByte(); } public short readShort() throws IOException { return reader.readShort(); } public int readInt() throws IOException { return reader.readInt(); } public long readLong() throws IOException { return reader.readLong(); } public float readFloat() throws IOException { return reader.readFloat(); } public double readDouble() throws IOException { return reader.readDouble(); } public byte[] readByteArray() throws IOException { int length = reader.readInt(); byte[] data = new byte[length]; reader.readBytes(data); return data; } public String readString() throws IOException { int length = reader.readShort(); byte[] data = new byte[length]; reader.readBytes(data); return new String(data, StandardCharsets.UTF_8); } @SuppressWarnings("unchecked") public NbtTagList<?> readList() throws IOException { byte type = reader.readByte(); int
length = reader.readInt();
NbtTagList<Object> list = new NbtTagList<>(type, length); for (int i = 0; i < length; i++) { list.add((NbtTagValue<Object>) readTag(type)); } return list; } public NbtTagCompound readCompound() throws IOException { NbtTagCompound compound = new NbtTagCompound(); for (byte type; (type = reader.readByte()) != NbtTagValue.TAG_END;) { compound.put(readString(), readTag(type)); } return compound; } public @Nullable NbtRootCompound readRootCompoundOptional() throws IOException { byte type = reader.readByte(); if (type == NbtTagValue.TAG_END) return null; return readRootCompoundInner(type); } public NbtRootCompound readRootCompound() throws IOException { return readRootCompoundInner(reader.readByte()); } private NbtRootCompound readRootCompoundInner(byte type) throws IOException { if (type != NbtTagValue.TAG_COMPOUND) { throw new RuntimeException("Root is not TAG_Compound type"); } return new NbtRootCompound(readString(), readCompound()); } public int[] readIntArray() throws IOException { int length = reader.readInt(); int[] value = new int[length]; for (int i = 0; i < length; i++) { value[i] = reader.readInt(); } return value; } public long[] readLongArray() throws IOException { int length = reader.readInt(); long[] value = new long[length]; for (int i = 0; i < length; i++) { value[i] = reader.readLong(); } return value; } public NbtTagValue<?> readTag(byte type) throws IOException { switch (type) { case NbtTagValue.TAG_BYTE: return NbtTagValue.fromByte(readByte()); case NbtTagValue.TAG_SHORT: return NbtTagValue.fromShort(readShort()); case NbtTagValue.TAG_INT: return NbtTagValue.fromInt(readInt()); case NbtTagValue.TAG_LONG: return NbtTagValue.fromLong(readLong()); case NbtTagValue.TAG_FLOAT: return NbtTagValue.fromFloat(readFloat()); case NbtTagValue.TAG_DOUBLE: return NbtTagValue.fromDouble(readDouble()); case NbtTagValue.TAG_BYTE_ARRAY: return NbtTagValue.fromByteArray(readByteArray()); case NbtTagValue.TAG_STRING: return NbtTagValue.fromString(readString()); case NbtTagValue.TAG_LIST: return NbtTagValue.fromList(readList()); case NbtTagValue.TAG_COMPOUND: return NbtTagValue.fromCompound(readCompound()); case NbtTagValue.TAG_INT_ARRAY: return NbtTagValue.fromIntArray(readIntArray()); case NbtTagValue.TAG_LONG_ARRAY: return NbtTagValue.fromLongArray(readLongArray()); default: throw new RuntimeException("Unknown NBT type: " + type); } } }
api/src/main/java/sh/pancake/serdemc/data/nbt/io/NbtReader.java
storycraft-serde-mc-757a9e4
[ { "filename": "api/src/main/java/sh/pancake/serdemc/network/io/PacketDataReader.java", "retrieved_chunk": " private final DataReader reader;\n private final NbtReader nbtReader;\n public PacketDataReader(DataReader reader) {\n this.reader = reader;\n this.nbtReader = new NbtReader(reader);\n }\n public byte readByte() throws IOException {\n return reader.readByte();\n }\n public int readByteUnsigned() throws IOException {", "score": 0.8666113018989563 }, { "filename": "api/src/main/java/sh/pancake/serdemc/network/io/PacketDataReader.java", "retrieved_chunk": " }\n public UUID readUUID() throws IOException {\n return new UUID(reader.readLong(), reader.readLong());\n }\n public @Nullable ItemStack readSlot() throws IOException {\n if (!readBoolean()) {\n return null;\n }\n return new ItemStack(readVarInt(), readByte(), readNbt());\n }", "score": 0.863715410232544 }, { "filename": "api/src/main/java/sh/pancake/serdemc/data/nbt/NbtTagList.java", "retrieved_chunk": " @Getter\n private final byte type;\n private final ArrayList<NbtTagValue<T>> list;\n public NbtTagList(byte type) {\n this.type = type;\n this.list = new ArrayList<>();\n }\n public NbtTagList(byte type, int length) {\n this.type = type;\n this.list = new ArrayList<>(length);", "score": 0.8624536991119385 }, { "filename": "api/src/main/java/sh/pancake/serdemc/network/io/PacketDataReader.java", "retrieved_chunk": " }\n public short readShort() throws IOException {\n return reader.readShort();\n }\n public int readShortUnsigned() throws IOException {\n return reader.readShortUnsigned();\n }\n public int readInt() throws IOException {\n return reader.readInt();\n }", "score": 0.8620169162750244 }, { "filename": "api/src/main/java/sh/pancake/serdemc/network/io/PacketDataReader.java", "retrieved_chunk": " }\n byte[] buf = new byte[length];\n reader.readBytes(buf);\n return new String(buf, StandardCharsets.UTF_8);\n }\n public String readChat() throws IOException {\n return readString(262144);\n }\n public String readIdentifier() throws IOException {\n return readString(32767);", "score": 0.8615350723266602 } ]
java
length = reader.readInt();
/* * Created on Sun Mar 26 2023 * * Copyright (c) storycraft. Licensed under the Apache Licence 2.0. */ package sh.pancake.serdemc.data.nbt.io; import java.io.IOException; import java.nio.charset.StandardCharsets; import javax.annotation.Nullable; import lombok.AllArgsConstructor; import sh.pancake.serdemc.data.nbt.NbtRootCompound; import sh.pancake.serdemc.data.nbt.NbtTagCompound; import sh.pancake.serdemc.data.nbt.NbtTagList; import sh.pancake.serdemc.data.nbt.NbtTagValue; import sh.pancake.serdemc.io.DataReader; @AllArgsConstructor public class NbtReader { private final DataReader reader; public byte readByte() throws IOException { return reader.readByte(); } public short readShort() throws IOException { return reader.readShort(); } public int readInt() throws IOException { return reader.readInt(); } public long readLong() throws IOException { return reader.readLong(); } public float readFloat() throws IOException { return reader.readFloat(); } public double readDouble() throws IOException { return reader.readDouble(); } public byte[] readByteArray() throws IOException { int length = reader.readInt(); byte[] data = new byte[length]; reader.readBytes(data); return data; } public String readString() throws IOException { int length = reader.readShort(); byte[] data = new byte[length]; reader.readBytes(data); return new String(data, StandardCharsets.UTF_8); } @SuppressWarnings("unchecked") public NbtTagList<?> readList() throws IOException { byte type = reader.readByte(); int length = reader.readInt(); NbtTagList<Object> list = new NbtTagList<>(type, length); for (int i = 0; i < length; i++) { list.add((NbtTagValue<Object>) readTag(type)); } return list; } public NbtTagCompound readCompound() throws IOException { NbtTagCompound compound = new NbtTagCompound(); for (byte type; (
type = reader.readByte()) != NbtTagValue.TAG_END;
) { compound.put(readString(), readTag(type)); } return compound; } public @Nullable NbtRootCompound readRootCompoundOptional() throws IOException { byte type = reader.readByte(); if (type == NbtTagValue.TAG_END) return null; return readRootCompoundInner(type); } public NbtRootCompound readRootCompound() throws IOException { return readRootCompoundInner(reader.readByte()); } private NbtRootCompound readRootCompoundInner(byte type) throws IOException { if (type != NbtTagValue.TAG_COMPOUND) { throw new RuntimeException("Root is not TAG_Compound type"); } return new NbtRootCompound(readString(), readCompound()); } public int[] readIntArray() throws IOException { int length = reader.readInt(); int[] value = new int[length]; for (int i = 0; i < length; i++) { value[i] = reader.readInt(); } return value; } public long[] readLongArray() throws IOException { int length = reader.readInt(); long[] value = new long[length]; for (int i = 0; i < length; i++) { value[i] = reader.readLong(); } return value; } public NbtTagValue<?> readTag(byte type) throws IOException { switch (type) { case NbtTagValue.TAG_BYTE: return NbtTagValue.fromByte(readByte()); case NbtTagValue.TAG_SHORT: return NbtTagValue.fromShort(readShort()); case NbtTagValue.TAG_INT: return NbtTagValue.fromInt(readInt()); case NbtTagValue.TAG_LONG: return NbtTagValue.fromLong(readLong()); case NbtTagValue.TAG_FLOAT: return NbtTagValue.fromFloat(readFloat()); case NbtTagValue.TAG_DOUBLE: return NbtTagValue.fromDouble(readDouble()); case NbtTagValue.TAG_BYTE_ARRAY: return NbtTagValue.fromByteArray(readByteArray()); case NbtTagValue.TAG_STRING: return NbtTagValue.fromString(readString()); case NbtTagValue.TAG_LIST: return NbtTagValue.fromList(readList()); case NbtTagValue.TAG_COMPOUND: return NbtTagValue.fromCompound(readCompound()); case NbtTagValue.TAG_INT_ARRAY: return NbtTagValue.fromIntArray(readIntArray()); case NbtTagValue.TAG_LONG_ARRAY: return NbtTagValue.fromLongArray(readLongArray()); default: throw new RuntimeException("Unknown NBT type: " + type); } } }
api/src/main/java/sh/pancake/serdemc/data/nbt/io/NbtReader.java
storycraft-serde-mc-757a9e4
[ { "filename": "api/src/main/java/sh/pancake/serdemc/network/io/PacketDataReader.java", "retrieved_chunk": " }\n public @Nullable NbtRootCompound readNbtOptional() throws IOException {\n return nbtReader.readRootCompoundOptional();\n }\n public Map<Byte, MetadataValue> readEntityMetadata() throws IOException {\n Map<Byte, MetadataValue> map = new HashMap<>();\n for (int index; (index = readByteUnsigned()) != 0xFF;) {\n int type = readVarInt();\n MetadataCodec<?> codec = MetadataCodec.getCodec(type);\n if (codec == null) {", "score": 0.8781821727752686 }, { "filename": "api/src/main/java/sh/pancake/serdemc/data/nbt/NbtTagList.java", "retrieved_chunk": " @Getter\n private final byte type;\n private final ArrayList<NbtTagValue<T>> list;\n public NbtTagList(byte type) {\n this.type = type;\n this.list = new ArrayList<>();\n }\n public NbtTagList(byte type, int length) {\n this.type = type;\n this.list = new ArrayList<>(length);", "score": 0.8758884072303772 }, { "filename": "api/src/main/java/sh/pancake/serdemc/data/nbt/io/NbtWriter.java", "retrieved_chunk": " writer.writeInt(length);\n for (int i = 0; i < length; i++) {\n writeLong(value[i]);\n }\n }\n @SuppressWarnings(\"unchecked\")\n public void writeTag(NbtTagValue<?> tag) throws IOException {\n byte type = tag.getType();\n Object value = tag.getValue();\n switch (type) {", "score": 0.8605108261108398 }, { "filename": "api/src/main/java/sh/pancake/serdemc/network/io/PacketDataReader.java", "retrieved_chunk": " public BlockPosition readPosition() throws IOException {\n long pos = reader.readLong();\n return new BlockPosition(\n (int) (pos >>> 38L),\n (int) ((pos >>> 12L) & 67108863L),\n (int) (pos & 4095L)\n );\n }\n public NbtRootCompound readNbt() throws IOException {\n return nbtReader.readRootCompound();", "score": 0.8560457229614258 }, { "filename": "api/src/main/java/sh/pancake/serdemc/data/nbt/io/NbtWriter.java", "retrieved_chunk": " writer.writeShortUnsigned(value.length());\n writer.writeBytes(value.getBytes());\n }\n public <Item> void writeList(NbtTagList<Item> list) throws IOException {\n writer.writeByte(list.getType());\n writer.writeInt(list.size());\n for (NbtTagValue<?> tag : list) {\n writeTag(tag);\n }\n }", "score": 0.8475595712661743 } ]
java
type = reader.readByte()) != NbtTagValue.TAG_END;
/* * Created on Sun Mar 26 2023 * * Copyright (c) storycraft. Licensed under the Apache Licence 2.0. */ package sh.pancake.serdemc.network.io; import java.io.IOException; import java.nio.charset.StandardCharsets; import java.util.HashMap; import java.util.Map; import java.util.UUID; import javax.annotation.Nullable; import sh.pancake.serdemc.data.BlockPosition; import sh.pancake.serdemc.data.ItemStack; import sh.pancake.serdemc.data.metadata.MetadataValue; import sh.pancake.serdemc.data.metadata.codec.MetadataCodec; import sh.pancake.serdemc.data.nbt.NbtRootCompound; import sh.pancake.serdemc.data.nbt.io.NbtReader; import sh.pancake.serdemc.io.DataReader; public class PacketDataReader { private final DataReader reader; private final NbtReader nbtReader; public PacketDataReader(DataReader reader) { this.reader = reader; this.nbtReader = new NbtReader(reader); } public byte readByte() throws IOException { return reader.readByte(); } public int readByteUnsigned() throws IOException { return reader.readByteUnsigned(); } public boolean readBoolean() throws IOException { if (reader.readByte() == 0x00) { return false; } return true; } public int readBytes(byte[] dst) throws IOException { return reader.readBytes(dst); } public short readShort() throws IOException { return reader.readShort(); } public int readShortUnsigned() throws IOException { return reader.readShortUnsigned(); } public int readInt() throws IOException { return reader.readInt(); } public int readVarInt() throws IOException { int value = 0; for (int position = 0;; position += 7) { if (position >= 32) throw new RuntimeException("VarInt is too big"); byte current = readByte(); value |= (current & 0x7F) << position; if ((current & 0x80) == 0) { break; } } return value; } public long readLong() throws IOException { return reader.readLong(); } public long readVarLong() throws IOException { long value = 0; for (int position = 0;; position += 7) { if (position >= 64) throw new RuntimeException("VarLong is too big"); byte current = readByte(); value |= (long) (current & 0x7F) << position; if ((current & 0x80) == 0) { break; } } return value; } public float readFloat() throws IOException { return reader.readFloat(); } public double readDouble() throws IOException { return reader.readDouble(); } public String readString() throws IOException { return readString(32767); } public String readString(int maxLength) throws IOException { int length = readVarInt(); if (length > maxLength) { throw new RuntimeException("String exceed max length. maxLength: " + maxLength + " length: " + length); } byte[] buf = new byte[length]; reader.readBytes(buf); return new String(buf, StandardCharsets.UTF_8); } public String readChat() throws IOException { return readString(262144); } public String readIdentifier() throws IOException { return readString(32767); } public UUID readUUID() throws IOException { return new UUID(reader.readLong(), reader.readLong()); } public @Nullable ItemStack readSlot() throws IOException { if (!readBoolean()) { return null; } return new ItemStack(readVarInt(), readByte(), readNbt()); } public BlockPosition readPosition() throws IOException { long pos = reader.readLong(); return new BlockPosition( (int) (pos >>> 38L), (int) ((pos >>> 12L) & 67108863L), (int) (pos & 4095L) ); } public NbtRootCompound readNbt() throws IOException { return nbtReader.readRootCompound(); } public @Nullable NbtRootCompound readNbtOptional() throws IOException { return nbtReader.readRootCompoundOptional(); } public Map<Byte, MetadataValue> readEntityMetadata() throws IOException { Map<Byte, MetadataValue> map = new HashMap<>(); for (int index; (index = readByteUnsigned()) != 0xFF;) { int type = readVarInt(); MetadataCodec<
?> codec = MetadataCodec.getCodec(type);
if (codec == null) { throw new RuntimeException("Unknown metadata type: " + type); } map.put((byte) index, new MetadataValue(type, codec.read(this))); } return map; } }
api/src/main/java/sh/pancake/serdemc/network/io/PacketDataReader.java
storycraft-serde-mc-757a9e4
[ { "filename": "api/src/main/java/sh/pancake/serdemc/network/io/PacketDataWriter.java", "retrieved_chunk": " }\n public void readNbtOptional(@Nullable NbtRootCompound compound) throws IOException {\n nbtWriter.writeRootCompoundOptional(compound);\n }\n @SuppressWarnings(\"unchecked\")\n public void writeEntityMetadata(Map<Byte, MetadataValue> metadata) throws IOException {\n for (Byte index : metadata.keySet()) {\n if (index.byteValue() == (byte) 0xFF) {\n throw new RuntimeException(\"Index cannot be 255\");\n }", "score": 0.9219729900360107 }, { "filename": "api/src/main/java/sh/pancake/serdemc/data/nbt/io/NbtReader.java", "retrieved_chunk": " for (byte type; (type = reader.readByte()) != NbtTagValue.TAG_END;) {\n compound.put(readString(), readTag(type));\n }\n return compound;\n }\n public @Nullable NbtRootCompound readRootCompoundOptional() throws IOException {\n byte type = reader.readByte();\n if (type == NbtTagValue.TAG_END) return null;\n return readRootCompoundInner(type);\n }", "score": 0.9213850498199463 }, { "filename": "api/src/main/java/sh/pancake/serdemc/data/nbt/io/NbtReader.java", "retrieved_chunk": " public NbtRootCompound readRootCompound() throws IOException {\n return readRootCompoundInner(reader.readByte());\n }\n private NbtRootCompound readRootCompoundInner(byte type) throws IOException {\n if (type != NbtTagValue.TAG_COMPOUND) {\n throw new RuntimeException(\"Root is not TAG_Compound type\");\n }\n return new NbtRootCompound(readString(), readCompound());\n }\n public int[] readIntArray() throws IOException {", "score": 0.8957035541534424 }, { "filename": "api/src/main/java/sh/pancake/serdemc/data/nbt/io/NbtReader.java", "retrieved_chunk": " byte type = reader.readByte();\n int length = reader.readInt();\n NbtTagList<Object> list = new NbtTagList<>(type, length);\n for (int i = 0; i < length; i++) {\n list.add((NbtTagValue<Object>) readTag(type));\n }\n return list;\n }\n public NbtTagCompound readCompound() throws IOException {\n NbtTagCompound compound = new NbtTagCompound();", "score": 0.8915359973907471 }, { "filename": "api/src/main/java/sh/pancake/serdemc/data/nbt/io/NbtReader.java", "retrieved_chunk": "import sh.pancake.serdemc.data.nbt.NbtRootCompound;\nimport sh.pancake.serdemc.data.nbt.NbtTagCompound;\nimport sh.pancake.serdemc.data.nbt.NbtTagList;\nimport sh.pancake.serdemc.data.nbt.NbtTagValue;\nimport sh.pancake.serdemc.io.DataReader;\n@AllArgsConstructor\npublic class NbtReader {\n private final DataReader reader;\n public byte readByte() throws IOException {\n return reader.readByte();", "score": 0.8834116458892822 } ]
java
?> codec = MetadataCodec.getCodec(type);
/* * Created on Sun Mar 26 2023 * * Copyright (c) storycraft. Licensed under the Apache Licence 2.0. */ package sh.pancake.serdemc.network.io; import java.io.IOException; import java.util.Map; import java.util.UUID; import javax.annotation.Nullable; import sh.pancake.serdemc.data.BlockPosition; import sh.pancake.serdemc.data.ItemStack; import sh.pancake.serdemc.data.metadata.MetadataValue; import sh.pancake.serdemc.data.metadata.codec.MetadataCodec; import sh.pancake.serdemc.data.nbt.NbtRootCompound; import sh.pancake.serdemc.data.nbt.io.NbtWriter; import sh.pancake.serdemc.io.DataWriter; public class PacketDataWriter { private final DataWriter writer; private final NbtWriter nbtWriter; public PacketDataWriter(DataWriter writer) { this.writer = writer; this.nbtWriter = new NbtWriter(writer); } public void writeByte(byte value) throws IOException { writer.writeByte(value); } public void writeByteUnsigned(int value) throws IOException { writer.writeByteUnsigned(value); } public void writeBoolean(boolean value) throws IOException { writer.writeByte((byte) (value ? 1 : 0)); } public void writeBytes(byte[] dst) throws IOException { writer.writeBytes(dst); } public void writeShort(short value) throws IOException { writer.writeShort(value); } public void writeShortUnsigned(int value) throws IOException { writer.writeShortUnsigned(value); } public void writeInt(int value) throws IOException { writer.writeInt(value); } public void writeVarInt(int value) throws IOException { while (true) { if ((value & ~0x7F) == 0) { writeByte((byte) value); return; } writeByte((byte) ((value & 0x7F) | 0x80)); value >>>= 7; } } public void writeLong(long value) throws IOException { writer.writeLong(value); } public void writeVarLong(long value) throws IOException { while (true) { if ((value & ~0x7FL) == 0) { writeByte((byte) value); return; } writeByte((byte) ((value & 0x7F) | 0x80)); value >>>= 7; } } public void writeFloat(float value) throws IOException { writer.writeFloat(value); } public void writeDouble(double value) throws IOException { writer.writeDouble(value); } public void writeString(String value) throws IOException { writeString(value, 32767); } public void writeString(String value, int maxLength) throws IOException { int length = value.length(); if (length > maxLength) { throw new RuntimeException("String exceed max length. maxLength: " + maxLength + " length: " + length); } writeVarInt(length); writeBytes(value.getBytes()); } public void writeChat(String value) throws IOException { writeString(value, 262144); } public void writeIdentifier(String value) throws IOException { writeString(value, 32767); } public void writeUUID(UUID uuid) throws IOException { writer.writeLong(uuid.getMostSignificantBits()); writer.writeLong(uuid.getLeastSignificantBits()); } public void writeSlot(@Nullable ItemStack item) throws IOException { if (item == null) { writeBoolean(false); return; } writeBoolean(true); writeVarInt(item.getId()); writeByte(item.getCount()); writeNbt(item.getNbt()); } public void writePosition(BlockPosition position) throws IOException { writer.writeLong(position.getX() << 38L | (position.getZ() & 0x3FFFFFFL) << 12L | position.getY() & 0xFFFL); } public void writeNbt(NbtRootCompound compound) throws IOException { nbtWriter.writeRootCompound(compound); } public void readNbtOptional(@Nullable NbtRootCompound compound) throws IOException { nbtWriter.writeRootCompoundOptional(compound); } @SuppressWarnings("unchecked") public void writeEntityMetadata(Map<Byte, MetadataValue> metadata) throws IOException { for (Byte index : metadata.keySet()) { if (index.byteValue() == (byte) 0xFF) { throw new RuntimeException("Index cannot be 255"); } MetadataValue value = metadata.get(index); int type = value.getType(); MetadataCodec<Object
> codec = (MetadataCodec<Object>) MetadataCodec.getCodec(type);
if (codec == null) { throw new RuntimeException("Unknown metadata type: " + type); } writeVarInt(type); codec.write(this, value.getValue()); } writeByte((byte) 0xFF); } }
api/src/main/java/sh/pancake/serdemc/network/io/PacketDataWriter.java
storycraft-serde-mc-757a9e4
[ { "filename": "api/src/main/java/sh/pancake/serdemc/network/io/PacketDataReader.java", "retrieved_chunk": " throw new RuntimeException(\"Unknown metadata type: \" + type);\n }\n map.put((byte) index, new MetadataValue(type, codec.read(this)));\n }\n return map;\n }\n}", "score": 0.8576703071594238 }, { "filename": "api/src/main/java/sh/pancake/serdemc/data/metadata/codec/MetadataCodec.java", "retrieved_chunk": " writer.writeFloat(value.getX());\n writer.writeFloat(value.getY());\n writer.writeFloat(value.getZ());\n writer.writeFloat(value.getW());\n }\n };\n private static final MetadataCodec<?>[] MAP = {\n BYTE, // 0 byte\n VAR_INT, // 1 varint\n VAR_LONG, // 2 varlong", "score": 0.8353212475776672 }, { "filename": "api/src/main/java/sh/pancake/serdemc/network/io/PacketDataReader.java", "retrieved_chunk": " }\n public @Nullable NbtRootCompound readNbtOptional() throws IOException {\n return nbtReader.readRootCompoundOptional();\n }\n public Map<Byte, MetadataValue> readEntityMetadata() throws IOException {\n Map<Byte, MetadataValue> map = new HashMap<>();\n for (int index; (index = readByteUnsigned()) != 0xFF;) {\n int type = readVarInt();\n MetadataCodec<?> codec = MetadataCodec.getCodec(type);\n if (codec == null) {", "score": 0.8301313519477844 }, { "filename": "api/src/main/java/sh/pancake/serdemc/data/metadata/codec/MetadataCodec.java", "retrieved_chunk": "public abstract class MetadataCodec<T> implements MetadataValueReader<T>, MetadataValueWriter<T> {\n public static final MetadataCodec<Byte> BYTE = create(PacketDataReader::readByte, PacketDataWriter::writeByte);\n public static final MetadataCodec<Integer> VAR_INT = create(PacketDataReader::readVarInt, PacketDataWriter::writeVarInt);\n public static final MetadataCodec<Long> VAR_LONG = create(PacketDataReader::readVarLong, PacketDataWriter::writeVarLong);\n public static final MetadataCodec<Float> FLOAT = create(PacketDataReader::readFloat, PacketDataWriter::writeFloat);\n public static final MetadataCodec<String> STRING = create(PacketDataReader::readString, PacketDataWriter::writeString);\n public static final MetadataCodec<String> CHAT = create(PacketDataReader::readChat, PacketDataWriter::writeChat);\n public static final MetadataCodec<ItemStack> SLOT = create(PacketDataReader::readSlot, PacketDataWriter::writeSlot);\n public static final MetadataCodec<Boolean> BOOLEAN = create(PacketDataReader::readBoolean, PacketDataWriter::writeBoolean);\n public static final MetadataCodec<BlockPosition> POSITION = create(PacketDataReader::readPosition, PacketDataWriter::writePosition);", "score": 0.8240719437599182 }, { "filename": "api/src/main/java/sh/pancake/serdemc/data/nbt/io/NbtWriter.java", "retrieved_chunk": " writer.writeInt(length);\n for (int i = 0; i < length; i++) {\n writeLong(value[i]);\n }\n }\n @SuppressWarnings(\"unchecked\")\n public void writeTag(NbtTagValue<?> tag) throws IOException {\n byte type = tag.getType();\n Object value = tag.getValue();\n switch (type) {", "score": 0.8218170404434204 } ]
java
> codec = (MetadataCodec<Object>) MetadataCodec.getCodec(type);
package com.github.stupdit1t.jackson.expand.cache; import com.github.stupdit1t.jackson.expand.config.JacksonExpandProperties; import com.github.stupdit1t.jackson.expand.util.SpringUtil; import com.github.stupdit1t.jackson.expand.serializer.ExpandSerializer; import org.springframework.util.AntPathMatcher; import java.time.Duration; import java.util.Set; import java.util.StringJoiner; /** * 缓存抽象 */ public interface ExpandCache { /** * 放入缓存 * * @param key * @param value * @param <T> */ <T> void put(String key, T value, Duration timeout); /** * 获取缓存 * * @param key * @return */ <T> T get(String key); /** * 列出匹配的的key * * @param pattern * @return */ Set<String> keys(String pattern); /** * 清空缓存 */ void clear(); /** * 删除缓存 * * @param key */ void delete(String key); /** * 按照bean删除缓存 */ default void delete(String beanName, String method, Object bindData, Object... annotationVal) { JacksonExpandProperties properties = SpringUtil.getBean(JacksonExpandProperties.class); StringJoiner key = new StringJoiner("-"); key.add(String.valueOf(bindData)); for (Object subVal : annotationVal) { key.add(String.valueOf(subVal)); } String cacheKey =
properties.getCachePrefix() + ":" + beanName + ":" + method + ":%s:" + key.toString();
delete(String.format(cacheKey, ExpandSerializer.OK)); delete(String.format(cacheKey, ExpandSerializer.FAIL)); } /** * 模糊匹配key * * @param pattern * @param key */ default boolean matchKey(String pattern, String key) { AntPathMatcher antPathMatcher = new AntPathMatcher(); // * if ("*".equals(pattern)) { return true; } // h?llo if (pattern.contains("?")) { if (antPathMatcher.match(pattern, key)) { return true; } } // h*llo if (pattern.contains("*")) { if (antPathMatcher.match(pattern, key)) { return true; } } // h[ae]llo if (pattern.contains("[") && pattern.contains("]")) { return key.matches(pattern); } return false; } }
src/main/java/com/github/stupdit1t/jackson/expand/cache/ExpandCache.java
stupdit1t-jackson-expand-boot-starter-b492b74
[ { "filename": "src/main/java/com/github/stupdit1t/jackson/expand/serializer/ExpandSerializer.java", "retrieved_chunk": " }\n if (bindData == null || loadService == null) {\n gen.writeObject(bindData);\n return;\n }\n // 获取缓存KEY\n Object[] args = params.getRemoteParams();\n int argsLength = args == null ? 0 : args.length;\n String cacheKey = jacksonExpandProperties.getCachePrefix() + \":\" + beanName + \":\" + method + \":%s:\" + paramsHandler.getCacheKey(bindData, args);\n Object result = getCacheInfo(cacheKey);", "score": 0.8466076850891113 }, { "filename": "src/main/java/com/github/stupdit1t/jackson/expand/handler/params/ParamsHandler.java", "retrieved_chunk": " */\n SerializerParam handleAnnotation(BeanProperty property);\n /**\n * 获取缓存的key\n *\n * @param val 当前值\n * @param annotationVal 注解值\n * @return\n */\n default String getCacheKey(Object val, Object[] annotationVal) {", "score": 0.8319376707077026 }, { "filename": "src/main/java/com/github/stupdit1t/jackson/expand/serializer/ExpandSerializer.java", "retrieved_chunk": " cache.put(String.format(cacheKey, OK), result, Duration.ofSeconds(cacheTime));\n } else {\n LOG.error(\"【{}】 Expand失败,未找到:{}\", beanName, bindData);\n cache.put(String.format(cacheKey, FAIL), bindData, Duration.ofSeconds(cacheTime));\n result = bindData;\n }\n } catch (Exception e) {\n LOG.error(\"【{}】 Expand异常:\", beanName, e);\n cache.put(String.format(cacheKey, FAIL), bindData, Duration.ofSeconds(cacheTime));\n result = bindData;", "score": 0.8150944709777832 }, { "filename": "src/main/java/com/github/stupdit1t/jackson/expand/serializer/ExpandSerializer.java", "retrieved_chunk": " this();\n this.loadService = SpringUtil.getBean(beanName);\n this.method = method;\n this.params = params;\n this.responseHandler = otherResponseHandler;\n this.paramsHandler = paramsHandler;\n this.beanName = beanName;\n }\n @Override\n public void serialize(Object bindData, JsonGenerator gen, SerializerProvider serializers) throws IOException {", "score": 0.8063814043998718 }, { "filename": "src/main/java/com/github/stupdit1t/jackson/expand/handler/params/ParamsHandler.java", "retrieved_chunk": " if (annotationVal == null) {\n return val.toString();\n }\n StringJoiner key = new StringJoiner(\"-\");\n key.add(String.valueOf(val));\n for (Object subVal : annotationVal) {\n key.add(String.valueOf(subVal));\n }\n return key.toString();\n }", "score": 0.8041313886642456 } ]
java
properties.getCachePrefix() + ":" + beanName + ":" + method + ":%s:" + key.toString();
package com.github.stupdit1t.jackson.expand.serializer; import com.fasterxml.jackson.core.JsonGenerator; import com.fasterxml.jackson.core.JsonStreamContext; import com.fasterxml.jackson.databind.BeanProperty; import com.fasterxml.jackson.databind.JsonMappingException; import com.fasterxml.jackson.databind.JsonSerializer; import com.fasterxml.jackson.databind.SerializerProvider; import com.fasterxml.jackson.databind.ser.ContextualSerializer; import com.github.stupdit1t.jackson.expand.annotation.Expand; import com.github.stupdit1t.jackson.expand.cache.ExpandCache; import com.github.stupdit1t.jackson.expand.cache.LocalCache; import com.github.stupdit1t.jackson.expand.config.JacksonExpandProperties; import com.github.stupdit1t.jackson.expand.domain.ExpandStrategy; import com.github.stupdit1t.jackson.expand.domain.SerializerParam; import com.github.stupdit1t.jackson.expand.handler.params.ParamsHandler; import com.github.stupdit1t.jackson.expand.handler.rsp.ResponseHandler; import com.github.stupdit1t.jackson.expand.util.ReflectUtil; import com.github.stupdit1t.jackson.expand.util.SpringUtil; import org.slf4j.Logger; import org.slf4j.LoggerFactory; import org.springframework.util.ReflectionUtils; import org.springframework.util.StringUtils; import org.springframework.web.context.request.RequestAttributes; import org.springframework.web.context.request.RequestContextHolder; import org.springframework.web.context.request.ServletRequestAttributes; import javax.servlet.ServletRequest; import java.io.IOException; import java.lang.reflect.Field; import java.time.Duration; import java.util.*; import java.util.concurrent.locks.StampedLock; import java.util.stream.Collectors; public class ExpandSerializer extends JsonSerializer<Object> implements ContextualSerializer { private static final Logger LOG = LoggerFactory.getLogger(ExpandSerializer.class); /** * 成功数据 */ public static final String OK = "OK"; /** * 失败数据 */ public static final String FAIL = "FAIL"; /** * 缓存 */ private static ExpandCache cache; /** * 配置 */ private static JacksonExpandProperties jacksonExpandProperties; /** * 本地锁缓存,防止同时查询 */ private static final LocalCache lockCache = new LocalCache(); /** * 远程调用服务 */ private Object loadService; /** * 方法 */ private String method; /** * 注解参数处理 */ private SerializerParam params; /** * 返回结果处理类 */ private ParamsHandler paramsHandler; /** * 返回结果处理类 */ private ResponseHandler responseHandler; /** * bean名称 */ private String beanName; public ExpandSerializer() { super(); if (cache == null) { synchronized (ExpandSerializer.class) { if (cache == null) { cache = SpringUtil.getBean(ExpandCache.class); jacksonExpandProperties = SpringUtil.getBean(JacksonExpandProperties.class); } } } } public ExpandSerializer(String beanName, String method, SerializerParam params, ParamsHandler paramsHandler, ResponseHandler otherResponseHandler) { this(); this.loadService = SpringUtil.getBean(beanName); this.method = method; this.params = params; this.responseHandler = otherResponseHandler; this.paramsHandler = paramsHandler; this.beanName = beanName; } @Override public void serialize(Object bindData, JsonGenerator gen, SerializerProvider serializers) throws IOException { String writeFieldPath = getFieldPath(gen.getOutputContext()); // 统一path替换 String dynamicExpandCommonPrefix = jacksonExpandProperties.getDynamicExpandCommonPrefix(); if (StringUtils.hasText(dynamicExpandCommonPrefix) && writeFieldPath.startsWith(dynamicExpandCommonPrefix)) { writeFieldPath = writeFieldPath.substring(dynamicExpandCommonPrefix.length() + 1); } // 是否展开 boolean expand; // 动态展开开启,判断是否展开 boolean dynamicExpand = jacksonExpandProperties.isDynamicExpand(); if (dynamicExpand) { Set<String> needExpandField = getParam(jacksonExpandProperties.getDynamicExpandParameterName()); // 如果代码里设置不展开,动态展开也不生效 expand = needExpandField.contains(writeFieldPath) && params.isOpen(); } else { expand = params.isOpen(); } if (!expand) { gen.writeObject(bindData); return; } // 判断要写入的字段 String writeField = gen.getOutputContext().getCurrentName(); if (jacksonExpandProperties.getExpandStrategy() == ExpandStrategy.COVER) { writeField = gen.getOutputContext().getCurrentName(); } else if (jacksonExpandProperties.getExpandStrategy() == ExpandStrategy.COPY) { writeField = String.format(jacksonExpandProperties.getCopyStrategyFormat(), gen.getOutputContext().getCurrentName()); } // 自定义要写入的优先级最高 if (StringUtils.hasText(params.getWriteField())) { writeField = params.getWriteField(); } // 设置理论上的响应类型,要不要使用取决于 ResponseHandler 要不要处理,比如只能写入数据对象存在的对象,默认是忽略存不存在 Class<?> writeClass = null; if (params.getWriteField() != null && StringUtils.hasText(params.getWriteField())) { Field field = ReflectionUtils.findField(gen.getCurrentValue().getClass(), params.getWriteField()); if (field != null) { writeClass = field.getType(); } } // 关闭不存在字段扩展,被写入的字段类型找不到,不扩展 if (!jacksonExpandProperties.isCanExpandToNotExistField() && writeClass == null) { gen.writeObject(bindData); return; } // 翻译为非当前字段,先写入当前字段值再翻译 boolean currField = gen.getOutputContext().getCurrentName().equals(writeField); if (!currField) { gen.writeObject(bindData); gen.writeFieldName(writeField); } if (bindData == null || loadService == null) { gen.writeObject(bindData); return; } // 获取缓存KEY Object[] args = params.getRemoteParams(); int argsLength = args == null ? 0 : args.length; String cacheKey = jacksonExpandProperties.getCachePrefix() + ":" + beanName + ":" + method + ":%s:" + paramsHandler.getCacheKey(bindData, args); Object result = getCacheInfo(cacheKey); if (result != null) { LOG.info("{} Expand cache 命中: {}", beanName, result); gen.writeObject(result); return; } StampedLock lock = lockCache.get(cacheKey, new StampedLock(), Duration.ofSeconds(300)); // 写锁避免同一业务ID重复查询 long stamp = lock.writeLock(); Integer cacheTime = params.getCacheTime(); try { // 多参数组装 Object[] objectParams = new Object[argsLength + 1]; objectParams[0] = paramsHandler.handleVal(bindData); if(objectParams.length > 1){ System.arraycopy(args, 0, objectParams, 1, argsLength); } // 请求翻译结果 Object loadResult = ReflectUtil.invoke(loadService, method, objectParams); if (loadResult != null) { result = this.responseHandler.handle(this.beanName, method, loadResult, writeClass, objectParams); cache.put(String.format(cacheKey, OK), result, Duration.ofSeconds(cacheTime)); } else { LOG.error("【{}】 Expand失败,未找到:{}", beanName, bindData);
cache.put(String.format(cacheKey, FAIL), bindData, Duration.ofSeconds(cacheTime));
result = bindData; } } catch (Exception e) { LOG.error("【{}】 Expand异常:", beanName, e); cache.put(String.format(cacheKey, FAIL), bindData, Duration.ofSeconds(cacheTime)); result = bindData; } finally { lock.unlockWrite(stamp); } gen.writeObject(result); } /** * 获取当前字段的path路径 * * @param outputContext * @return */ private String getFieldPath(JsonStreamContext outputContext) { List<String> path = new ArrayList<>(4); while (outputContext != null) { String currentName = outputContext.getCurrentName(); if (StringUtils.hasText(currentName)) { path.add(currentName); } outputContext = outputContext.getParent(); } Collections.reverse(path); return String.join(".", path); } /** * 获取厍信息 * * @param cacheKey 缓存的KEY * @return */ private Object getCacheInfo(String cacheKey) { Object result = cache.get(String.format(cacheKey, OK)); if (result == null) { result = cache.get(String.format(cacheKey, FAIL)); } return result; } @Override public JsonSerializer<?> createContextual(SerializerProvider prov, BeanProperty property) throws JsonMappingException { if (property != null) { Expand load = property.getAnnotation(Expand.class); if (load == null) { throw new RuntimeException("未注解相关 @Expand 注解"); } String bean = load.bean(); Class<? extends ParamsHandler> paramsHandlerClass = load.paramsHandler(); Class<? extends ResponseHandler> responseHandlerClass = load.responseHandler(); String method = load.method(); try { ParamsHandler paramsHandler = paramsHandlerClass.getDeclaredConstructor().newInstance(); ResponseHandler responseHandler = responseHandlerClass.getDeclaredConstructor().newInstance(); int cacheTime = load.cacheTime(); // 额外参数处理 SerializerParam params = paramsHandler.handleAnnotation(property); // 参数处理器没设置,且父注设置了,以父注解为主 if (params.getCacheTime() == null && cacheTime != -1) { params.setCacheTime(cacheTime); } // 缓存时间未设置,取默认 if (params.getCacheTime() == null) { params.setCacheTime(jacksonExpandProperties.getCacheTimeout()); } if (params.isOpen() == null) { params.setExpand(load.expand()); } return new ExpandSerializer(bean, method, params, paramsHandler, responseHandler); } catch (Exception e) { LOG.error("@Expand error: ", e); } } return prov.findNullValueSerializer(null); } /** * 获取展开参数 * * @param key * @return */ private Set<String> getParam(String key) { RequestAttributes attributes = RequestContextHolder.getRequestAttributes(); if (attributes == null) { return Collections.emptySet(); } ServletRequest request = ((ServletRequestAttributes) attributes).getRequest(); String[] parameterValues = request.getParameterValues(key); if (parameterValues == null) { return Collections.emptySet(); } return Arrays.stream(parameterValues).flatMap(o -> Arrays.stream(o.split(","))) .collect(Collectors.toSet()); } }
src/main/java/com/github/stupdit1t/jackson/expand/serializer/ExpandSerializer.java
stupdit1t-jackson-expand-boot-starter-b492b74
[ { "filename": "src/main/java/com/github/stupdit1t/jackson/expand/cache/ExpandCache.java", "retrieved_chunk": " key.add(String.valueOf(bindData));\n for (Object subVal : annotationVal) {\n key.add(String.valueOf(subVal));\n }\n String cacheKey = properties.getCachePrefix() + \":\" + beanName + \":\" + method + \":%s:\" + key.toString();\n delete(String.format(cacheKey, ExpandSerializer.OK));\n delete(String.format(cacheKey, ExpandSerializer.FAIL));\n }\n /**\n * 模糊匹配key", "score": 0.8275107741355896 }, { "filename": "src/main/java/com/github/stupdit1t/jackson/expand/handler/rsp/DefaultResponseHandler.java", "retrieved_chunk": " * @param rsp 当前返回值\n * @param toClass 要填充字段的类型\n * @param params 当前方法参数\n * @return\n */\n @Override\n public Object handle(String bean, String method, Object rsp, Class<?> toClass, Object... params) {\n return rsp;\n }\n}", "score": 0.8150525093078613 }, { "filename": "src/main/java/com/github/stupdit1t/jackson/expand/cache/ExpandCache.java", "retrieved_chunk": " *\n * @param key\n */\n void delete(String key);\n /**\n * 按照bean删除缓存\n */\n default void delete(String beanName, String method, Object bindData, Object... annotationVal) {\n JacksonExpandProperties properties = SpringUtil.getBean(JacksonExpandProperties.class);\n StringJoiner key = new StringJoiner(\"-\");", "score": 0.8057459592819214 }, { "filename": "src/main/java/com/github/stupdit1t/jackson/expand/handler/params/DefaultParamsHandler.java", "retrieved_chunk": " if (expand != null) {\n if (StringUtils.hasText(expand.to())) {\n params.setWriteField(expand.to());\n }\n params.setExpand(expand.expand());\n }\n return params;\n }\n}", "score": 0.8022830486297607 }, { "filename": "src/main/java/com/github/stupdit1t/jackson/expand/handler/rsp/ResponseHandler.java", "retrieved_chunk": "package com.github.stupdit1t.jackson.expand.handler.rsp;\n/**\n * 返回处理第三方响应\n */\npublic interface ResponseHandler {\n Object handle(String bean, String method, Object rsp, Class<?> writeClass, Object... params);\n}", "score": 0.801891028881073 } ]
java
cache.put(String.format(cacheKey, FAIL), bindData, Duration.ofSeconds(cacheTime));
/* * Copyright Strimzi authors. * License: Apache License 2.0 (see the file LICENSE or http://apache.org/licenses/LICENSE-2.0.html). */ package io.strimzi.kafka.bridge.mqtt.mapper; import java.util.HashMap; import java.util.List; import java.util.Map; import java.util.regex.Matcher; import java.util.regex.Pattern; /** * Responsible for handling all the topic mapping using named placeholders instead of regular expressions. * * @see MappingRule * @see MqttKafkaMapper * @see MqttKafkaRegexMapper */ public class MqttKafkaSimpleMapper extends MqttKafkaMapper { // find any word inside a curly bracket. E.g. {something}, this is known as a placeholder. private static final String MQTT_TOPIC_PLACEHOLDER_REGEX = "\\{\\w+\\}"; // identifies a single level wildcard character in the mqtt pattern. E.g. sensors/+/data private static final String MQTT_TOPIC_SINGLE_LEVEL_WILDCARD_CHARACTER = "+"; // Regex expression used to replace the + in the mqtt pattern. private static final String SINGLE_LEVEL_WILDCARD_REGEX = "[^/]+"; // identifies a multi level wildcard character in the mqtt pattern. E.g. sensors/# private static final String MQTT_TOPIC_MULTI_LEVEL_WILDCARD_CHARACTER = "#"; // used to replace the # in the mqtt pattern. public static final String WILDCARD_REGEX = "(?:\\/.*)?$"; /** * Constructor. * * @param rules the list of mapping rules. */ public MqttKafkaSimpleMapper(List<MappingRule> rules) { super(rules, Pattern.compile(MQTT_TOPIC_PLACEHOLDER_REGEX)); } @Override public MappingResult map(String mqttTopic) { for (MappingRule rule : this.rules) { Matcher matcher = this.patterns.get(this.rules.indexOf(rule)).matcher(mqttTopic); if (matcher.matches()) { HashMap<String, String> placeholders = new HashMap<>(); String mappedKafkaTopic = rule.getKafkaTopicTemplate(); String kafkaKey = rule.getKafkaKeyTemplate(); // find MQTT_TOPIC_PLACEHOLDER_REGEX in the kafkaTopicTemplate. Matcher placeholderMatcher = this.placeholderPattern.matcher(rule.getKafkaTopicTemplate()); while (placeholderMatcher.find()) { String placeholderKey = placeholderMatcher.group(); placeholders.put(placeholderKey, null); } // find MQTT_TOPIC_PLACEHOLDER_REGEX in the kafkaKey if (kafkaKey != null) { placeholderMatcher = this.placeholderPattern.matcher(kafkaKey); while (placeholderMatcher.find()) { String placeholderKey = placeholderMatcher.group(); placeholders.put(placeholderKey, null); } } if (!placeholders.isEmpty()) { Matcher mqttTopicMatcher = this.placeholderPattern.
matcher(rule.getMqttTopicPattern());
// find the placeholders in the mqtt topic pattern and assign them a value. while (mqttTopicMatcher.find()) { String placeholderKey = mqttTopicMatcher.group(); String placeholderValue = matcher.group(removeBrackets(placeholderKey)); placeholders.put(placeholderKey, placeholderValue); } // build the Kafka topic using the placeholders. for (Map.Entry<String, String> entry : placeholders.entrySet()) { if (entry.getValue() != null) { mappedKafkaTopic = mappedKafkaTopic.replace(entry.getKey(), entry.getValue()); kafkaKey = kafkaKey != null ? kafkaKey.replace(entry.getKey(), entry.getValue()) : null; } else { throw new IllegalArgumentException("The placeholder " + entry.getKey() + " was not found assigned any value."); } } } return new MappingResult(mappedKafkaTopic, kafkaKey); } } return new MappingResult(MqttKafkaMapper.DEFAULT_KAFKA_TOPIC, null); } @Override protected void buildOrCompilePatterns() { // convert the mqtt patterns to a valid regex expression. // the mqtt pattern can contain placeholders like {something}, + and #. // if the mqtt topic contains a +, we replace it with @singleLevelWildcardRegex // if the mqtt topic contains a #, we replace it with @multiLevelWildcardRegex // if the mqtt topic contains a placeholder (pattern \{\w+\}), we replace it with @placeholderRegex String[] mqttTopicPatternParts; StringBuilder ruleRegex; for (MappingRule rule : this.rules) { mqttTopicPatternParts = rule.getMqttTopicPattern().split(MQTT_TOPIC_SEPARATOR); ruleRegex = new StringBuilder(); for (String part : mqttTopicPatternParts) { if (part.matches(MQTT_TOPIC_PLACEHOLDER_REGEX)) { ruleRegex.append(buildNamedRegexExpression(part)); } else if (part.equals(MQTT_TOPIC_SINGLE_LEVEL_WILDCARD_CHARACTER)) { ruleRegex.append(SINGLE_LEVEL_WILDCARD_REGEX); } else if (part.equals(MQTT_TOPIC_MULTI_LEVEL_WILDCARD_CHARACTER)) { if (ruleRegex.length() > 1) { ruleRegex.deleteCharAt(ruleRegex.length() - 1); } ruleRegex.append(WILDCARD_REGEX); } else { ruleRegex.append(part); } ruleRegex.append(MQTT_TOPIC_SEPARATOR); } // remove the last slash ruleRegex.deleteCharAt(ruleRegex.length() - 1); // compile the regex expression for the rule. patterns.add(Pattern.compile(ruleRegex.toString())); } } /** * Helper method for building a named regex expression. * A named regex expression is a regex expression that contains a named capturing group. * E.g. (?<groupName>regexExpression) * * @param placeholder represents a placeholder in the mqtt pattern. * @return a named regex expression. */ private String buildNamedRegexExpression(String placeholder) { String groupName = removeBrackets(placeholder); return "(?<" + groupName + ">[^/]+)"; } /** * Helper method for removing the curly brackets from a placeholder. * * @param placeholder represents a placeholder in the pattern. * @return a placeholder without the curly brackets. */ private String removeBrackets(String placeholder) { return placeholder.replaceAll("\\{+|\\}+", ""); } }
src/main/java/io/strimzi/kafka/bridge/mqtt/mapper/MqttKafkaSimpleMapper.java
strimzi-strimzi-mqtt-bridge-cb0a4b8
[ { "filename": "src/main/java/io/strimzi/kafka/bridge/mqtt/mapper/MqttKafkaRegexMapper.java", "retrieved_chunk": " }\n }\n return new MappingResult(MqttKafkaMapper.DEFAULT_KAFKA_TOPIC, null);\n }\n @Override\n protected void buildOrCompilePatterns() {\n this.rules.forEach(rule-> this.patterns.add(Pattern.compile(rule.getMqttTopicPattern())));\n }\n /**\n * Checks if there are any pending placeholders in the Kafka topic or Kafka key template.", "score": 0.8935047388076782 }, { "filename": "src/main/java/io/strimzi/kafka/bridge/mqtt/mapper/MqttKafkaRegexMapper.java", "retrieved_chunk": " kafkaKey = kafkaKey != null ? kafkaKey.replace(\"$\" + i, matcher.group(i)) : null;\n }\n // check for pending placeholders replacement in the Kafka topic\n checkPlaceholder(mappedKafkaTopic);\n if (kafkaKey != null) {\n // check for pending placeholders replacement in the Kafka key.\n checkPlaceholder(kafkaKey);\n }\n // return the first match\n return new MappingResult(mappedKafkaTopic, kafkaKey);", "score": 0.8876934051513672 }, { "filename": "src/main/java/io/strimzi/kafka/bridge/mqtt/mapper/MqttKafkaRegexMapper.java", "retrieved_chunk": " }\n @Override\n public MappingResult map(String mqttTopic) {\n for (MappingRule rule : this.rules) {\n Matcher matcher = this.patterns.get(this.rules.indexOf(rule)).matcher(mqttTopic);\n if (matcher.matches()) {\n String mappedKafkaTopic = rule.getKafkaTopicTemplate();\n String kafkaKey = rule.getKafkaKeyTemplate();\n for (int i = 1; i < matcher.groupCount() + 1; i++) {\n mappedKafkaTopic = mappedKafkaTopic.replace(\"$\" + i, matcher.group(i));", "score": 0.8852992057800293 }, { "filename": "src/main/java/io/strimzi/kafka/bridge/mqtt/mapper/MqttKafkaMapper.java", "retrieved_chunk": " * Constructor\n *\n * @param rules the list of mapping rules.\n * @param placeholderPattern the pattern used to find placeholders.\n * @see MappingRule\n */\n protected MqttKafkaMapper(List<MappingRule> rules, Pattern placeholderPattern) {\n this.rules = rules;\n this.placeholderPattern = placeholderPattern;\n this.buildOrCompilePatterns();", "score": 0.859525203704834 }, { "filename": "src/main/java/io/strimzi/kafka/bridge/mqtt/mapper/MqttKafkaMapper.java", "retrieved_chunk": " */\npublic abstract class MqttKafkaMapper {\n // default Kafka topic. Used when no mapping rule matches the mqtt topic.\n public static final String DEFAULT_KAFKA_TOPIC = \"messages_default\";\n // MQTT topic separator\n public static final String MQTT_TOPIC_SEPARATOR = \"/\";\n protected final List<MappingRule> rules;\n protected final List<Pattern> patterns = new ArrayList<>();\n protected final Pattern placeholderPattern;\n /**", "score": 0.8583739995956421 } ]
java
matcher(rule.getMqttTopicPattern());
package com.github.stupdit1t.jackson.expand.serializer; import com.fasterxml.jackson.core.JsonGenerator; import com.fasterxml.jackson.core.JsonStreamContext; import com.fasterxml.jackson.databind.BeanProperty; import com.fasterxml.jackson.databind.JsonMappingException; import com.fasterxml.jackson.databind.JsonSerializer; import com.fasterxml.jackson.databind.SerializerProvider; import com.fasterxml.jackson.databind.ser.ContextualSerializer; import com.github.stupdit1t.jackson.expand.annotation.Expand; import com.github.stupdit1t.jackson.expand.cache.ExpandCache; import com.github.stupdit1t.jackson.expand.cache.LocalCache; import com.github.stupdit1t.jackson.expand.config.JacksonExpandProperties; import com.github.stupdit1t.jackson.expand.domain.ExpandStrategy; import com.github.stupdit1t.jackson.expand.domain.SerializerParam; import com.github.stupdit1t.jackson.expand.handler.params.ParamsHandler; import com.github.stupdit1t.jackson.expand.handler.rsp.ResponseHandler; import com.github.stupdit1t.jackson.expand.util.ReflectUtil; import com.github.stupdit1t.jackson.expand.util.SpringUtil; import org.slf4j.Logger; import org.slf4j.LoggerFactory; import org.springframework.util.ReflectionUtils; import org.springframework.util.StringUtils; import org.springframework.web.context.request.RequestAttributes; import org.springframework.web.context.request.RequestContextHolder; import org.springframework.web.context.request.ServletRequestAttributes; import javax.servlet.ServletRequest; import java.io.IOException; import java.lang.reflect.Field; import java.time.Duration; import java.util.*; import java.util.concurrent.locks.StampedLock; import java.util.stream.Collectors; public class ExpandSerializer extends JsonSerializer<Object> implements ContextualSerializer { private static final Logger LOG = LoggerFactory.getLogger(ExpandSerializer.class); /** * 成功数据 */ public static final String OK = "OK"; /** * 失败数据 */ public static final String FAIL = "FAIL"; /** * 缓存 */ private static ExpandCache cache; /** * 配置 */ private static JacksonExpandProperties jacksonExpandProperties; /** * 本地锁缓存,防止同时查询 */ private static final LocalCache lockCache = new LocalCache(); /** * 远程调用服务 */ private Object loadService; /** * 方法 */ private String method; /** * 注解参数处理 */ private SerializerParam params; /** * 返回结果处理类 */ private ParamsHandler paramsHandler; /** * 返回结果处理类 */ private ResponseHandler responseHandler; /** * bean名称 */ private String beanName; public ExpandSerializer() { super(); if (cache == null) { synchronized (ExpandSerializer.class) { if (cache == null) { cache = SpringUtil.getBean(ExpandCache.class); jacksonExpandProperties = SpringUtil.getBean(JacksonExpandProperties.class); } } } } public ExpandSerializer(String beanName, String method, SerializerParam params, ParamsHandler paramsHandler, ResponseHandler otherResponseHandler) { this(); this.loadService = SpringUtil.getBean(beanName); this.method = method; this.params = params; this.responseHandler = otherResponseHandler; this.paramsHandler = paramsHandler; this.beanName = beanName; } @Override public void serialize(Object bindData, JsonGenerator gen, SerializerProvider serializers) throws IOException { String writeFieldPath = getFieldPath(gen.getOutputContext()); // 统一path替换 String dynamicExpandCommonPrefix = jacksonExpandProperties.getDynamicExpandCommonPrefix(); if (StringUtils.hasText(dynamicExpandCommonPrefix) && writeFieldPath.startsWith(dynamicExpandCommonPrefix)) { writeFieldPath = writeFieldPath.substring(dynamicExpandCommonPrefix.length() + 1); } // 是否展开 boolean expand; // 动态展开开启,判断是否展开 boolean dynamicExpand = jacksonExpandProperties.isDynamicExpand(); if (dynamicExpand) { Set<String> needExpandField = getParam(jacksonExpandProperties.getDynamicExpandParameterName()); // 如果代码里设置不展开,动态展开也不生效 expand = needExpandField.contains(writeFieldPath) && params.isOpen(); } else { expand = params.isOpen(); } if (!expand) { gen.writeObject(bindData); return; } // 判断要写入的字段 String writeField = gen.getOutputContext().getCurrentName(); if (jacksonExpandProperties.getExpandStrategy() == ExpandStrategy.COVER) { writeField = gen.getOutputContext().getCurrentName(); } else if (jacksonExpandProperties.getExpandStrategy() == ExpandStrategy.COPY) { writeField = String.format(jacksonExpandProperties.getCopyStrategyFormat(), gen.getOutputContext().getCurrentName()); } // 自定义要写入的优先级最高 if (StringUtils.hasText(params.getWriteField())) { writeField = params.getWriteField(); } // 设置理论上的响应类型,要不要使用取决于 ResponseHandler 要不要处理,比如只能写入数据对象存在的对象,默认是忽略存不存在 Class<?> writeClass = null; if (params.getWriteField() != null && StringUtils.hasText(params.getWriteField())) { Field field = ReflectionUtils.findField(gen.getCurrentValue().getClass(), params.getWriteField()); if (field != null) { writeClass = field.getType(); } } // 关闭不存在字段扩展,被写入的字段类型找不到,不扩展 if
(!jacksonExpandProperties.isCanExpandToNotExistField() && writeClass == null) {
gen.writeObject(bindData); return; } // 翻译为非当前字段,先写入当前字段值再翻译 boolean currField = gen.getOutputContext().getCurrentName().equals(writeField); if (!currField) { gen.writeObject(bindData); gen.writeFieldName(writeField); } if (bindData == null || loadService == null) { gen.writeObject(bindData); return; } // 获取缓存KEY Object[] args = params.getRemoteParams(); int argsLength = args == null ? 0 : args.length; String cacheKey = jacksonExpandProperties.getCachePrefix() + ":" + beanName + ":" + method + ":%s:" + paramsHandler.getCacheKey(bindData, args); Object result = getCacheInfo(cacheKey); if (result != null) { LOG.info("{} Expand cache 命中: {}", beanName, result); gen.writeObject(result); return; } StampedLock lock = lockCache.get(cacheKey, new StampedLock(), Duration.ofSeconds(300)); // 写锁避免同一业务ID重复查询 long stamp = lock.writeLock(); Integer cacheTime = params.getCacheTime(); try { // 多参数组装 Object[] objectParams = new Object[argsLength + 1]; objectParams[0] = paramsHandler.handleVal(bindData); if(objectParams.length > 1){ System.arraycopy(args, 0, objectParams, 1, argsLength); } // 请求翻译结果 Object loadResult = ReflectUtil.invoke(loadService, method, objectParams); if (loadResult != null) { result = this.responseHandler.handle(this.beanName, method, loadResult, writeClass, objectParams); cache.put(String.format(cacheKey, OK), result, Duration.ofSeconds(cacheTime)); } else { LOG.error("【{}】 Expand失败,未找到:{}", beanName, bindData); cache.put(String.format(cacheKey, FAIL), bindData, Duration.ofSeconds(cacheTime)); result = bindData; } } catch (Exception e) { LOG.error("【{}】 Expand异常:", beanName, e); cache.put(String.format(cacheKey, FAIL), bindData, Duration.ofSeconds(cacheTime)); result = bindData; } finally { lock.unlockWrite(stamp); } gen.writeObject(result); } /** * 获取当前字段的path路径 * * @param outputContext * @return */ private String getFieldPath(JsonStreamContext outputContext) { List<String> path = new ArrayList<>(4); while (outputContext != null) { String currentName = outputContext.getCurrentName(); if (StringUtils.hasText(currentName)) { path.add(currentName); } outputContext = outputContext.getParent(); } Collections.reverse(path); return String.join(".", path); } /** * 获取厍信息 * * @param cacheKey 缓存的KEY * @return */ private Object getCacheInfo(String cacheKey) { Object result = cache.get(String.format(cacheKey, OK)); if (result == null) { result = cache.get(String.format(cacheKey, FAIL)); } return result; } @Override public JsonSerializer<?> createContextual(SerializerProvider prov, BeanProperty property) throws JsonMappingException { if (property != null) { Expand load = property.getAnnotation(Expand.class); if (load == null) { throw new RuntimeException("未注解相关 @Expand 注解"); } String bean = load.bean(); Class<? extends ParamsHandler> paramsHandlerClass = load.paramsHandler(); Class<? extends ResponseHandler> responseHandlerClass = load.responseHandler(); String method = load.method(); try { ParamsHandler paramsHandler = paramsHandlerClass.getDeclaredConstructor().newInstance(); ResponseHandler responseHandler = responseHandlerClass.getDeclaredConstructor().newInstance(); int cacheTime = load.cacheTime(); // 额外参数处理 SerializerParam params = paramsHandler.handleAnnotation(property); // 参数处理器没设置,且父注设置了,以父注解为主 if (params.getCacheTime() == null && cacheTime != -1) { params.setCacheTime(cacheTime); } // 缓存时间未设置,取默认 if (params.getCacheTime() == null) { params.setCacheTime(jacksonExpandProperties.getCacheTimeout()); } if (params.isOpen() == null) { params.setExpand(load.expand()); } return new ExpandSerializer(bean, method, params, paramsHandler, responseHandler); } catch (Exception e) { LOG.error("@Expand error: ", e); } } return prov.findNullValueSerializer(null); } /** * 获取展开参数 * * @param key * @return */ private Set<String> getParam(String key) { RequestAttributes attributes = RequestContextHolder.getRequestAttributes(); if (attributes == null) { return Collections.emptySet(); } ServletRequest request = ((ServletRequestAttributes) attributes).getRequest(); String[] parameterValues = request.getParameterValues(key); if (parameterValues == null) { return Collections.emptySet(); } return Arrays.stream(parameterValues).flatMap(o -> Arrays.stream(o.split(","))) .collect(Collectors.toSet()); } }
src/main/java/com/github/stupdit1t/jackson/expand/serializer/ExpandSerializer.java
stupdit1t-jackson-expand-boot-starter-b492b74
[ { "filename": "src/main/java/com/github/stupdit1t/jackson/expand/handler/params/DefaultParamsHandler.java", "retrieved_chunk": " if (expand != null) {\n if (StringUtils.hasText(expand.to())) {\n params.setWriteField(expand.to());\n }\n params.setExpand(expand.expand());\n }\n return params;\n }\n}", "score": 0.8686022758483887 }, { "filename": "src/main/java/com/github/stupdit1t/jackson/expand/domain/SerializerParam.java", "retrieved_chunk": " * 写入字段\n */\n private String writeField;\n /**\n * 值缓存时间\n */\n private Integer cacheTime;\n /**\n * 是否展开\n */", "score": 0.83758544921875 }, { "filename": "src/main/java/com/github/stupdit1t/jackson/expand/domain/SerializerParam.java", "retrieved_chunk": " private Boolean expand;\n public Object[] getRemoteParams() {\n return remoteParams;\n }\n public void setRemoteParams(Object[] remoteParams) {\n this.remoteParams = remoteParams;\n }\n public String getWriteField() {\n return writeField;\n }", "score": 0.8353731632232666 }, { "filename": "src/main/java/com/github/stupdit1t/jackson/expand/domain/SerializerParam.java", "retrieved_chunk": " public void setWriteField(String writeField) {\n this.writeField = writeField;\n }\n public Integer getCacheTime() {\n return cacheTime;\n }\n public void setCacheTime(Integer cacheTime) {\n this.cacheTime = cacheTime;\n }\n public Boolean isOpen() {", "score": 0.8017486929893494 }, { "filename": "src/main/java/com/github/stupdit1t/jackson/expand/handler/rsp/DefaultResponseHandler.java", "retrieved_chunk": " * @param rsp 当前返回值\n * @param toClass 要填充字段的类型\n * @param params 当前方法参数\n * @return\n */\n @Override\n public Object handle(String bean, String method, Object rsp, Class<?> toClass, Object... params) {\n return rsp;\n }\n}", "score": 0.790462851524353 } ]
java
(!jacksonExpandProperties.isCanExpandToNotExistField() && writeClass == null) {
package com.github.stupdit1t.jackson.expand.serializer; import com.fasterxml.jackson.core.JsonGenerator; import com.fasterxml.jackson.core.JsonStreamContext; import com.fasterxml.jackson.databind.BeanProperty; import com.fasterxml.jackson.databind.JsonMappingException; import com.fasterxml.jackson.databind.JsonSerializer; import com.fasterxml.jackson.databind.SerializerProvider; import com.fasterxml.jackson.databind.ser.ContextualSerializer; import com.github.stupdit1t.jackson.expand.annotation.Expand; import com.github.stupdit1t.jackson.expand.cache.ExpandCache; import com.github.stupdit1t.jackson.expand.cache.LocalCache; import com.github.stupdit1t.jackson.expand.config.JacksonExpandProperties; import com.github.stupdit1t.jackson.expand.domain.ExpandStrategy; import com.github.stupdit1t.jackson.expand.domain.SerializerParam; import com.github.stupdit1t.jackson.expand.handler.params.ParamsHandler; import com.github.stupdit1t.jackson.expand.handler.rsp.ResponseHandler; import com.github.stupdit1t.jackson.expand.util.ReflectUtil; import com.github.stupdit1t.jackson.expand.util.SpringUtil; import org.slf4j.Logger; import org.slf4j.LoggerFactory; import org.springframework.util.ReflectionUtils; import org.springframework.util.StringUtils; import org.springframework.web.context.request.RequestAttributes; import org.springframework.web.context.request.RequestContextHolder; import org.springframework.web.context.request.ServletRequestAttributes; import javax.servlet.ServletRequest; import java.io.IOException; import java.lang.reflect.Field; import java.time.Duration; import java.util.*; import java.util.concurrent.locks.StampedLock; import java.util.stream.Collectors; public class ExpandSerializer extends JsonSerializer<Object> implements ContextualSerializer { private static final Logger LOG = LoggerFactory.getLogger(ExpandSerializer.class); /** * 成功数据 */ public static final String OK = "OK"; /** * 失败数据 */ public static final String FAIL = "FAIL"; /** * 缓存 */ private static ExpandCache cache; /** * 配置 */ private static JacksonExpandProperties jacksonExpandProperties; /** * 本地锁缓存,防止同时查询 */ private static final LocalCache lockCache = new LocalCache(); /** * 远程调用服务 */ private Object loadService; /** * 方法 */ private String method; /** * 注解参数处理 */ private SerializerParam params; /** * 返回结果处理类 */ private ParamsHandler paramsHandler; /** * 返回结果处理类 */ private ResponseHandler responseHandler; /** * bean名称 */ private String beanName; public ExpandSerializer() { super(); if (cache == null) { synchronized (ExpandSerializer.class) { if (cache == null) { cache = SpringUtil.getBean(ExpandCache.class); jacksonExpandProperties = SpringUtil.getBean(JacksonExpandProperties.class); } } } } public ExpandSerializer(String beanName, String method, SerializerParam params, ParamsHandler paramsHandler, ResponseHandler otherResponseHandler) { this(); this.loadService = SpringUtil.getBean(beanName); this.method = method; this.params = params; this.responseHandler = otherResponseHandler; this.paramsHandler = paramsHandler; this.beanName = beanName; } @Override public void serialize(Object bindData, JsonGenerator gen, SerializerProvider serializers) throws IOException { String writeFieldPath = getFieldPath(gen.getOutputContext()); // 统一path替换 String dynamicExpandCommonPrefix = jacksonExpandProperties.getDynamicExpandCommonPrefix(); if (StringUtils.hasText(dynamicExpandCommonPrefix) && writeFieldPath.startsWith(dynamicExpandCommonPrefix)) { writeFieldPath = writeFieldPath.substring(dynamicExpandCommonPrefix.length() + 1); } // 是否展开 boolean expand; // 动态展开开启,判断是否展开 boolean dynamicExpand = jacksonExpandProperties.isDynamicExpand(); if (dynamicExpand) { Set<String> needExpandField = getParam(jacksonExpandProperties.getDynamicExpandParameterName()); // 如果代码里设置不展开,动态展开也不生效 expand = needExpandField.contains(writeFieldPath) && params.isOpen(); } else { expand = params.isOpen(); } if (!expand) { gen.writeObject(bindData); return; } // 判断要写入的字段 String writeField = gen.getOutputContext().getCurrentName(); if (jacksonExpandProperties.getExpandStrategy() == ExpandStrategy.COVER) { writeField = gen.getOutputContext().getCurrentName(); } else
if (jacksonExpandProperties.getExpandStrategy() == ExpandStrategy.COPY) {
writeField = String.format(jacksonExpandProperties.getCopyStrategyFormat(), gen.getOutputContext().getCurrentName()); } // 自定义要写入的优先级最高 if (StringUtils.hasText(params.getWriteField())) { writeField = params.getWriteField(); } // 设置理论上的响应类型,要不要使用取决于 ResponseHandler 要不要处理,比如只能写入数据对象存在的对象,默认是忽略存不存在 Class<?> writeClass = null; if (params.getWriteField() != null && StringUtils.hasText(params.getWriteField())) { Field field = ReflectionUtils.findField(gen.getCurrentValue().getClass(), params.getWriteField()); if (field != null) { writeClass = field.getType(); } } // 关闭不存在字段扩展,被写入的字段类型找不到,不扩展 if (!jacksonExpandProperties.isCanExpandToNotExistField() && writeClass == null) { gen.writeObject(bindData); return; } // 翻译为非当前字段,先写入当前字段值再翻译 boolean currField = gen.getOutputContext().getCurrentName().equals(writeField); if (!currField) { gen.writeObject(bindData); gen.writeFieldName(writeField); } if (bindData == null || loadService == null) { gen.writeObject(bindData); return; } // 获取缓存KEY Object[] args = params.getRemoteParams(); int argsLength = args == null ? 0 : args.length; String cacheKey = jacksonExpandProperties.getCachePrefix() + ":" + beanName + ":" + method + ":%s:" + paramsHandler.getCacheKey(bindData, args); Object result = getCacheInfo(cacheKey); if (result != null) { LOG.info("{} Expand cache 命中: {}", beanName, result); gen.writeObject(result); return; } StampedLock lock = lockCache.get(cacheKey, new StampedLock(), Duration.ofSeconds(300)); // 写锁避免同一业务ID重复查询 long stamp = lock.writeLock(); Integer cacheTime = params.getCacheTime(); try { // 多参数组装 Object[] objectParams = new Object[argsLength + 1]; objectParams[0] = paramsHandler.handleVal(bindData); if(objectParams.length > 1){ System.arraycopy(args, 0, objectParams, 1, argsLength); } // 请求翻译结果 Object loadResult = ReflectUtil.invoke(loadService, method, objectParams); if (loadResult != null) { result = this.responseHandler.handle(this.beanName, method, loadResult, writeClass, objectParams); cache.put(String.format(cacheKey, OK), result, Duration.ofSeconds(cacheTime)); } else { LOG.error("【{}】 Expand失败,未找到:{}", beanName, bindData); cache.put(String.format(cacheKey, FAIL), bindData, Duration.ofSeconds(cacheTime)); result = bindData; } } catch (Exception e) { LOG.error("【{}】 Expand异常:", beanName, e); cache.put(String.format(cacheKey, FAIL), bindData, Duration.ofSeconds(cacheTime)); result = bindData; } finally { lock.unlockWrite(stamp); } gen.writeObject(result); } /** * 获取当前字段的path路径 * * @param outputContext * @return */ private String getFieldPath(JsonStreamContext outputContext) { List<String> path = new ArrayList<>(4); while (outputContext != null) { String currentName = outputContext.getCurrentName(); if (StringUtils.hasText(currentName)) { path.add(currentName); } outputContext = outputContext.getParent(); } Collections.reverse(path); return String.join(".", path); } /** * 获取厍信息 * * @param cacheKey 缓存的KEY * @return */ private Object getCacheInfo(String cacheKey) { Object result = cache.get(String.format(cacheKey, OK)); if (result == null) { result = cache.get(String.format(cacheKey, FAIL)); } return result; } @Override public JsonSerializer<?> createContextual(SerializerProvider prov, BeanProperty property) throws JsonMappingException { if (property != null) { Expand load = property.getAnnotation(Expand.class); if (load == null) { throw new RuntimeException("未注解相关 @Expand 注解"); } String bean = load.bean(); Class<? extends ParamsHandler> paramsHandlerClass = load.paramsHandler(); Class<? extends ResponseHandler> responseHandlerClass = load.responseHandler(); String method = load.method(); try { ParamsHandler paramsHandler = paramsHandlerClass.getDeclaredConstructor().newInstance(); ResponseHandler responseHandler = responseHandlerClass.getDeclaredConstructor().newInstance(); int cacheTime = load.cacheTime(); // 额外参数处理 SerializerParam params = paramsHandler.handleAnnotation(property); // 参数处理器没设置,且父注设置了,以父注解为主 if (params.getCacheTime() == null && cacheTime != -1) { params.setCacheTime(cacheTime); } // 缓存时间未设置,取默认 if (params.getCacheTime() == null) { params.setCacheTime(jacksonExpandProperties.getCacheTimeout()); } if (params.isOpen() == null) { params.setExpand(load.expand()); } return new ExpandSerializer(bean, method, params, paramsHandler, responseHandler); } catch (Exception e) { LOG.error("@Expand error: ", e); } } return prov.findNullValueSerializer(null); } /** * 获取展开参数 * * @param key * @return */ private Set<String> getParam(String key) { RequestAttributes attributes = RequestContextHolder.getRequestAttributes(); if (attributes == null) { return Collections.emptySet(); } ServletRequest request = ((ServletRequestAttributes) attributes).getRequest(); String[] parameterValues = request.getParameterValues(key); if (parameterValues == null) { return Collections.emptySet(); } return Arrays.stream(parameterValues).flatMap(o -> Arrays.stream(o.split(","))) .collect(Collectors.toSet()); } }
src/main/java/com/github/stupdit1t/jackson/expand/serializer/ExpandSerializer.java
stupdit1t-jackson-expand-boot-starter-b492b74
[ { "filename": "src/main/java/com/github/stupdit1t/jackson/expand/handler/params/DefaultParamsHandler.java", "retrieved_chunk": " if (expand != null) {\n if (StringUtils.hasText(expand.to())) {\n params.setWriteField(expand.to());\n }\n params.setExpand(expand.expand());\n }\n return params;\n }\n}", "score": 0.8498305082321167 }, { "filename": "src/main/java/com/github/stupdit1t/jackson/expand/domain/SerializerParam.java", "retrieved_chunk": " * 写入字段\n */\n private String writeField;\n /**\n * 值缓存时间\n */\n private Integer cacheTime;\n /**\n * 是否展开\n */", "score": 0.8421859741210938 }, { "filename": "src/main/java/com/github/stupdit1t/jackson/expand/domain/ExpandStrategy.java", "retrieved_chunk": "package com.github.stupdit1t.jackson.expand.domain;\n/**\n * 展开策略\n */\npublic enum ExpandStrategy {\n /**\n * 覆盖当前字段\n */\n COVER,\n /**", "score": 0.840333104133606 }, { "filename": "src/main/java/com/github/stupdit1t/jackson/expand/config/JacksonExpandProperties.java", "retrieved_chunk": " */\n private String dynamicExpandParameterName = \"expand\";\n /**\n * 动态展开 统一数据的Path前缀,比如前缀是 data.body. 如果配置 expand=userId, 相当于是expnad=data.body.userId, 默认无\n */\n private String dynamicExpandCommonPrefix;\n /**\n * 展开策略, 默认覆盖\n */\n private ExpandStrategy expandStrategy = ExpandStrategy.COVER;", "score": 0.8261722922325134 }, { "filename": "src/main/java/com/github/stupdit1t/jackson/expand/domain/SerializerParam.java", "retrieved_chunk": " private Boolean expand;\n public Object[] getRemoteParams() {\n return remoteParams;\n }\n public void setRemoteParams(Object[] remoteParams) {\n this.remoteParams = remoteParams;\n }\n public String getWriteField() {\n return writeField;\n }", "score": 0.8242703080177307 } ]
java
if (jacksonExpandProperties.getExpandStrategy() == ExpandStrategy.COPY) {
/* * Copyright Strimzi authors. * License: Apache License 2.0 (see the file LICENSE or http://apache.org/licenses/LICENSE-2.0.html). */ package io.strimzi.kafka.bridge.mqtt.mapper; import java.util.HashMap; import java.util.List; import java.util.Map; import java.util.regex.Matcher; import java.util.regex.Pattern; /** * Responsible for handling all the topic mapping using named placeholders instead of regular expressions. * * @see MappingRule * @see MqttKafkaMapper * @see MqttKafkaRegexMapper */ public class MqttKafkaSimpleMapper extends MqttKafkaMapper { // find any word inside a curly bracket. E.g. {something}, this is known as a placeholder. private static final String MQTT_TOPIC_PLACEHOLDER_REGEX = "\\{\\w+\\}"; // identifies a single level wildcard character in the mqtt pattern. E.g. sensors/+/data private static final String MQTT_TOPIC_SINGLE_LEVEL_WILDCARD_CHARACTER = "+"; // Regex expression used to replace the + in the mqtt pattern. private static final String SINGLE_LEVEL_WILDCARD_REGEX = "[^/]+"; // identifies a multi level wildcard character in the mqtt pattern. E.g. sensors/# private static final String MQTT_TOPIC_MULTI_LEVEL_WILDCARD_CHARACTER = "#"; // used to replace the # in the mqtt pattern. public static final String WILDCARD_REGEX = "(?:\\/.*)?$"; /** * Constructor. * * @param rules the list of mapping rules. */ public MqttKafkaSimpleMapper(List<MappingRule> rules) { super(rules, Pattern.compile(MQTT_TOPIC_PLACEHOLDER_REGEX)); } @Override public MappingResult map(String mqttTopic) { for (MappingRule rule : this.rules) { Matcher matcher = this.patterns.get(this.rules.indexOf(rule)).matcher(mqttTopic); if (matcher.matches()) { HashMap<String, String> placeholders = new HashMap<>(); String mappedKafkaTopic = rule.getKafkaTopicTemplate(); String kafkaKey = rule.getKafkaKeyTemplate(); // find MQTT_TOPIC_PLACEHOLDER_REGEX in the kafkaTopicTemplate. Matcher placeholderMatcher = this.placeholderPattern.matcher(rule.getKafkaTopicTemplate()); while (placeholderMatcher.find()) { String placeholderKey = placeholderMatcher.group(); placeholders.put(placeholderKey, null); } // find MQTT_TOPIC_PLACEHOLDER_REGEX in the kafkaKey if (kafkaKey != null) { placeholderMatcher = this.placeholderPattern.matcher(kafkaKey); while (placeholderMatcher.find()) { String placeholderKey = placeholderMatcher.group(); placeholders.put(placeholderKey, null); } } if (!placeholders.isEmpty()) { Matcher mqttTopicMatcher = this.placeholderPattern.matcher(rule.getMqttTopicPattern()); // find the placeholders in the mqtt topic pattern and assign them a value. while (mqttTopicMatcher.find()) { String placeholderKey = mqttTopicMatcher.group(); String placeholderValue = matcher.group(removeBrackets(placeholderKey)); placeholders.put(placeholderKey, placeholderValue); } // build the Kafka topic using the placeholders. for (Map.Entry<String, String> entry : placeholders.entrySet()) { if (entry.getValue() != null) { mappedKafkaTopic = mappedKafkaTopic.replace(entry.getKey(), entry.getValue()); kafkaKey = kafkaKey != null ? kafkaKey.replace(entry.getKey(), entry.getValue()) : null; } else { throw new IllegalArgumentException("The placeholder " + entry.getKey() + " was not found assigned any value."); } } } return new MappingResult(mappedKafkaTopic, kafkaKey); } } return new MappingResult(MqttKafkaMapper.DEFAULT_KAFKA_TOPIC, null); } @Override protected void buildOrCompilePatterns() { // convert the mqtt patterns to a valid regex expression. // the mqtt pattern can contain placeholders like {something}, + and #. // if the mqtt topic contains a +, we replace it with @singleLevelWildcardRegex // if the mqtt topic contains a #, we replace it with @multiLevelWildcardRegex // if the mqtt topic contains a placeholder (pattern \{\w+\}), we replace it with @placeholderRegex String[] mqttTopicPatternParts; StringBuilder ruleRegex; for (MappingRule rule : this.rules) { mqttTopicPatternParts
= rule.getMqttTopicPattern().split(MQTT_TOPIC_SEPARATOR);
ruleRegex = new StringBuilder(); for (String part : mqttTopicPatternParts) { if (part.matches(MQTT_TOPIC_PLACEHOLDER_REGEX)) { ruleRegex.append(buildNamedRegexExpression(part)); } else if (part.equals(MQTT_TOPIC_SINGLE_LEVEL_WILDCARD_CHARACTER)) { ruleRegex.append(SINGLE_LEVEL_WILDCARD_REGEX); } else if (part.equals(MQTT_TOPIC_MULTI_LEVEL_WILDCARD_CHARACTER)) { if (ruleRegex.length() > 1) { ruleRegex.deleteCharAt(ruleRegex.length() - 1); } ruleRegex.append(WILDCARD_REGEX); } else { ruleRegex.append(part); } ruleRegex.append(MQTT_TOPIC_SEPARATOR); } // remove the last slash ruleRegex.deleteCharAt(ruleRegex.length() - 1); // compile the regex expression for the rule. patterns.add(Pattern.compile(ruleRegex.toString())); } } /** * Helper method for building a named regex expression. * A named regex expression is a regex expression that contains a named capturing group. * E.g. (?<groupName>regexExpression) * * @param placeholder represents a placeholder in the mqtt pattern. * @return a named regex expression. */ private String buildNamedRegexExpression(String placeholder) { String groupName = removeBrackets(placeholder); return "(?<" + groupName + ">[^/]+)"; } /** * Helper method for removing the curly brackets from a placeholder. * * @param placeholder represents a placeholder in the pattern. * @return a placeholder without the curly brackets. */ private String removeBrackets(String placeholder) { return placeholder.replaceAll("\\{+|\\}+", ""); } }
src/main/java/io/strimzi/kafka/bridge/mqtt/mapper/MqttKafkaSimpleMapper.java
strimzi-strimzi-mqtt-bridge-cb0a4b8
[ { "filename": "src/main/java/io/strimzi/kafka/bridge/mqtt/mapper/MqttKafkaRegexMapper.java", "retrieved_chunk": " }\n @Override\n public MappingResult map(String mqttTopic) {\n for (MappingRule rule : this.rules) {\n Matcher matcher = this.patterns.get(this.rules.indexOf(rule)).matcher(mqttTopic);\n if (matcher.matches()) {\n String mappedKafkaTopic = rule.getKafkaTopicTemplate();\n String kafkaKey = rule.getKafkaKeyTemplate();\n for (int i = 1; i < matcher.groupCount() + 1; i++) {\n mappedKafkaTopic = mappedKafkaTopic.replace(\"$\" + i, matcher.group(i));", "score": 0.8965745568275452 }, { "filename": "src/test/java/io/strimzi/kafka/bridge/mqtt/mapper/MqttKafkaRegexMapperTest.java", "retrieved_chunk": " rules.add(new MappingRule(\"([^/]+)/.*\", \"$1\", null));\n MqttKafkaRegexMapper mapper = new MqttKafkaRegexMapper(rules);\n // Test for building/([^/]+)/room/(\\\\d{1,3}).* pattern\n MappingResult mappingResult = mapper.map(\"building/4/room/23/temperature\");\n assertThat(\"Mqtt pattern building/([^/]+)/room/(\\\\d{1,3}).* should be mapped to building_$1\",\n mappingResult.kafkaTopic(), is(\"building_4\"));\n assertThat(\"The key for building/([^/]+)/room/(\\\\d{1,3}).* should be expanded to room_$2\",\n mappingResult.kafkaKey(), is(\"room_23\"));\n // Test for building/([^/]+).* pattern\n mappingResult = mapper.map(\"building/405/room\");", "score": 0.8690344095230103 }, { "filename": "src/main/java/io/strimzi/kafka/bridge/mqtt/mapper/MappingRule.java", "retrieved_chunk": " * \"mqttTopic\": \"sensors/(^[0-9])/type/([^/]+)/data\",\n * \"kafkaTopic\": \"sensors_$1_data\",\n * \"kafkaKey\": \"sensor_$2\"\n * }\n * and like this in the MappingRule class:\n * MappingRule(mqttTopicPattern=sensors/(^[0-9])/type/([^/]+)/data, kafkaTopicTemplate=sensors_$1_data, kafkaKey=sensor_$2)\n */\npublic class MappingRule {\n @JsonProperty(\"mqttTopic\")\n private String mqttTopicPattern;", "score": 0.8651536703109741 }, { "filename": "src/main/java/io/strimzi/kafka/bridge/mqtt/mapper/MqttKafkaMapper.java", "retrieved_chunk": " */\npublic abstract class MqttKafkaMapper {\n // default Kafka topic. Used when no mapping rule matches the mqtt topic.\n public static final String DEFAULT_KAFKA_TOPIC = \"messages_default\";\n // MQTT topic separator\n public static final String MQTT_TOPIC_SEPARATOR = \"/\";\n protected final List<MappingRule> rules;\n protected final List<Pattern> patterns = new ArrayList<>();\n protected final Pattern placeholderPattern;\n /**", "score": 0.8528246879577637 }, { "filename": "src/main/java/io/strimzi/kafka/bridge/mqtt/mapper/MqttKafkaRegexMapper.java", "retrieved_chunk": " */\npublic class MqttKafkaRegexMapper extends MqttKafkaMapper {\n // used to find any expression starting with a $ followed by upto 2 digits number. E.g. $1, this is known as a placeholder.\n public static final String MQTT_TOPIC_DOLLAR_PLACEHOLDER_REGEX = \"\\\\$(\\\\d{1,2})\";\n /**\n * Constructor\n * Creates a new instance of MqttKafkaRegexMapper.\n */\n public MqttKafkaRegexMapper(List<MappingRule> rules) {\n super(rules, Pattern.compile(MQTT_TOPIC_DOLLAR_PLACEHOLDER_REGEX));", "score": 0.8524303436279297 } ]
java
= rule.getMqttTopicPattern().split(MQTT_TOPIC_SEPARATOR);
package com.solodroid.ads.sdkdemo.activity; import android.os.Bundle; import android.view.MenuItem; import android.widget.Toast; import androidx.appcompat.app.ActionBar; import androidx.appcompat.app.AppCompatActivity; import androidx.appcompat.widget.Toolbar; import androidx.recyclerview.widget.DividerItemDecoration; import androidx.recyclerview.widget.RecyclerView; import androidx.recyclerview.widget.StaggeredGridLayoutManager; import com.solodroid.ads.sdk.format.BannerAd; import com.solodroid.ads.sdkdemo.R; import com.solodroid.ads.sdkdemo.adapter.AdapterPost; import com.solodroid.ads.sdkdemo.data.Constant; import com.solodroid.ads.sdkdemo.database.SharedPref; import com.solodroid.ads.sdkdemo.model.Post; import java.util.ArrayList; import java.util.List; public class SecondActivity extends AppCompatActivity { SharedPref sharedPref; RecyclerView recyclerView; AdapterPost adapterPost; BannerAd.Builder bannerAd; Toolbar toolbar; @Override protected void onCreate(Bundle savedInstanceState) { super.onCreate(savedInstanceState); sharedPref = new SharedPref(this); getAppTheme(); setContentView(R.layout.activity_second); initView(); loadBannerAd(); initToolbar(); } private void initView() { recyclerView = findViewById(R.id.recyclerView); recyclerView.setLayoutManager(new StaggeredGridLayoutManager(1, StaggeredGridLayoutManager.VERTICAL)); recyclerView.addItemDecoration(new DividerItemDecoration(this, DividerItemDecoration.VERTICAL)); adapterPost = new AdapterPost(this, new ArrayList<>()); recyclerView.setAdapter(adapterPost); displayData(sharedPref.getPostList()); } private void initToolbar() { toolbar = findViewById(R.id.toolbar); setSupportActionBar(toolbar); final ActionBar actionBar = getSupportActionBar(); if (actionBar != null) { getSupportActionBar().setDisplayHomeAsUpEnabled(true); getSupportActionBar().setHomeButtonEnabled(true); getSupportActionBar().setTitle("Second Activity"); } } private void displayData(List<Post> posts) { if (posts != null && posts.size() > 0) { adapterPost.setListData(posts, posts.size());
adapterPost.setOnItemClickListener((view, obj, position) -> {
Toast.makeText(getApplicationContext(), "" + obj.name, Toast.LENGTH_SHORT).show(); }); } } @Override public void onBackPressed() { super.onBackPressed(); bannerAd.destroyAndDetachBanner(); } private void loadBannerAd() { bannerAd = new BannerAd.Builder(this) .setAdStatus(Constant.AD_STATUS) .setAdNetwork(Constant.AD_NETWORK) .setBackupAdNetwork(Constant.BACKUP_AD_NETWORK) .setAdMobBannerId(Constant.ADMOB_BANNER_ID) .setGoogleAdManagerBannerId(Constant.GOOGLE_AD_MANAGER_BANNER_ID) .setUnityBannerId(Constant.UNITY_BANNER_ID) .setAppLovinBannerId(Constant.APPLOVIN_BANNER_ID) .setAppLovinBannerZoneId(Constant.APPLOVIN_BANNER_ZONE_ID) .setIronSourceBannerId(Constant.IRONSOURCE_BANNER_ID) .setWortiseBannerId(Constant.WORTISE_BANNER_ID) .setDarkTheme(false) .build(); } @Override public boolean onOptionsItemSelected(MenuItem menuItem) { if (menuItem.getItemId() == android.R.id.home) { onBackPressed(); return true; } return super.onOptionsItemSelected(menuItem); } public void getAppTheme() { if (sharedPref.getIsDarkTheme()) { setTheme(R.style.AppDarkTheme); } else { setTheme(R.style.AppTheme); } } }
demo/src/main/java/com/solodroid/ads/sdkdemo/activity/SecondActivity.java
solodroidx-multi-ads-sdk-4f23cdf
[ { "filename": "demo/src/main/java/com/solodroid/ads/sdkdemo/adapter/AdapterPost.java", "retrieved_chunk": " this.onItemClickListener = mItemClickListener;\n }\n public AdapterPost(Context context, List<Post> posts) {\n this.posts = posts;\n this.context = context;\n this.sharedPref = new SharedPref(context);\n }\n public static class OriginalViewHolder extends RecyclerView.ViewHolder {\n public TextView name;\n public ImageView image;", "score": 0.8403022289276123 }, { "filename": "demo/src/main/java/com/solodroid/ads/sdkdemo/adapter/AdapterPost.java", "retrieved_chunk": " @Override\n public int getItemCount() {\n return posts.size();\n }\n}", "score": 0.835783839225769 }, { "filename": "demo/src/main/java/com/solodroid/ads/sdkdemo/adapter/AdapterPost.java", "retrieved_chunk": "import com.bumptech.glide.Glide;\nimport com.bumptech.glide.load.engine.DiskCacheStrategy;\nimport com.solodroid.ads.sdk.format.NativeAdViewHolder;\nimport com.solodroid.ads.sdkdemo.R;\nimport com.solodroid.ads.sdkdemo.data.Constant;\nimport com.solodroid.ads.sdkdemo.database.SharedPref;\nimport com.solodroid.ads.sdkdemo.model.Post;\nimport java.util.List;\n@SuppressLint(\"NotifyDataSetChanged\")\npublic class AdapterPost extends RecyclerView.Adapter<RecyclerView.ViewHolder> {", "score": 0.8319941759109497 }, { "filename": "demo/src/main/java/com/solodroid/ads/sdkdemo/adapter/AdapterPost.java", "retrieved_chunk": " }\n @Override\n public void onBindViewHolder(@NonNull RecyclerView.ViewHolder holder, final int position) {\n if (holder instanceof OriginalViewHolder) {\n final Post p = posts.get(position);\n OriginalViewHolder vItem = (OriginalViewHolder) holder;\n vItem.name.setText(p.name);\n Glide.with(context)\n .load(p.image.replace(\" \", \"%20\"))\n .thumbnail(0.1f)", "score": 0.8311768174171448 }, { "filename": "demo/src/main/java/com/solodroid/ads/sdkdemo/adapter/AdapterPost.java", "retrieved_chunk": " }\n }\n public void setListData(List<Post> posts, int totalPosts) {\n this.posts = posts;\n insertNativeAd(posts, totalPosts);\n notifyDataSetChanged();\n }\n private void insertNativeAd(List<Post> posts, int totalPosts) {\n switch (Constant.AD_NETWORK) {\n case ADMOB:", "score": 0.8274437785148621 } ]
java
adapterPost.setOnItemClickListener((view, obj, position) -> {
package com.github.stupdit1t.jackson.expand.serializer; import com.fasterxml.jackson.core.JsonGenerator; import com.fasterxml.jackson.core.JsonStreamContext; import com.fasterxml.jackson.databind.BeanProperty; import com.fasterxml.jackson.databind.JsonMappingException; import com.fasterxml.jackson.databind.JsonSerializer; import com.fasterxml.jackson.databind.SerializerProvider; import com.fasterxml.jackson.databind.ser.ContextualSerializer; import com.github.stupdit1t.jackson.expand.annotation.Expand; import com.github.stupdit1t.jackson.expand.cache.ExpandCache; import com.github.stupdit1t.jackson.expand.cache.LocalCache; import com.github.stupdit1t.jackson.expand.config.JacksonExpandProperties; import com.github.stupdit1t.jackson.expand.domain.ExpandStrategy; import com.github.stupdit1t.jackson.expand.domain.SerializerParam; import com.github.stupdit1t.jackson.expand.handler.params.ParamsHandler; import com.github.stupdit1t.jackson.expand.handler.rsp.ResponseHandler; import com.github.stupdit1t.jackson.expand.util.ReflectUtil; import com.github.stupdit1t.jackson.expand.util.SpringUtil; import org.slf4j.Logger; import org.slf4j.LoggerFactory; import org.springframework.util.ReflectionUtils; import org.springframework.util.StringUtils; import org.springframework.web.context.request.RequestAttributes; import org.springframework.web.context.request.RequestContextHolder; import org.springframework.web.context.request.ServletRequestAttributes; import javax.servlet.ServletRequest; import java.io.IOException; import java.lang.reflect.Field; import java.time.Duration; import java.util.*; import java.util.concurrent.locks.StampedLock; import java.util.stream.Collectors; public class ExpandSerializer extends JsonSerializer<Object> implements ContextualSerializer { private static final Logger LOG = LoggerFactory.getLogger(ExpandSerializer.class); /** * 成功数据 */ public static final String OK = "OK"; /** * 失败数据 */ public static final String FAIL = "FAIL"; /** * 缓存 */ private static ExpandCache cache; /** * 配置 */ private static JacksonExpandProperties jacksonExpandProperties; /** * 本地锁缓存,防止同时查询 */ private static final LocalCache lockCache = new LocalCache(); /** * 远程调用服务 */ private Object loadService; /** * 方法 */ private String method; /** * 注解参数处理 */ private SerializerParam params; /** * 返回结果处理类 */ private ParamsHandler paramsHandler; /** * 返回结果处理类 */ private ResponseHandler responseHandler; /** * bean名称 */ private String beanName; public ExpandSerializer() { super(); if (cache == null) { synchronized (ExpandSerializer.class) { if (cache == null) { cache = SpringUtil.getBean(ExpandCache.class); jacksonExpandProperties = SpringUtil.getBean(JacksonExpandProperties.class); } } } } public ExpandSerializer(String beanName, String method, SerializerParam params, ParamsHandler paramsHandler, ResponseHandler otherResponseHandler) { this(); this.loadService = SpringUtil.getBean(beanName); this.method = method; this.params = params; this.responseHandler = otherResponseHandler; this.paramsHandler = paramsHandler; this.beanName = beanName; } @Override public void serialize(Object bindData, JsonGenerator gen, SerializerProvider serializers) throws IOException { String writeFieldPath = getFieldPath(gen.getOutputContext()); // 统一path替换 String dynamicExpandCommonPrefix = jacksonExpandProperties.getDynamicExpandCommonPrefix(); if (StringUtils.hasText(dynamicExpandCommonPrefix) && writeFieldPath.startsWith(dynamicExpandCommonPrefix)) { writeFieldPath = writeFieldPath.substring(dynamicExpandCommonPrefix.length() + 1); } // 是否展开 boolean expand; // 动态展开开启,判断是否展开 boolean dynamicExpand = jacksonExpandProperties.isDynamicExpand(); if (dynamicExpand) { Set<String> needExpandField = getParam(jacksonExpandProperties.getDynamicExpandParameterName()); // 如果代码里设置不展开,动态展开也不生效 expand = needExpandField.contains(writeFieldPath) && params.isOpen(); } else { expand = params.isOpen(); } if (!expand) { gen.writeObject(bindData); return; } // 判断要写入的字段 String writeField = gen.getOutputContext().getCurrentName(); if (jacksonExpandProperties.getExpandStrategy() == ExpandStrategy.COVER) { writeField = gen.getOutputContext().getCurrentName(); } else if (jacksonExpandProperties.getExpandStrategy() == ExpandStrategy.COPY) { writeField = String.format(jacksonExpandProperties.getCopyStrategyFormat(), gen.getOutputContext().getCurrentName()); } // 自定义要写入的优先级最高 if (StringUtils.hasText(params.getWriteField())) { writeField = params.getWriteField(); } // 设置理论上的响应类型,要不要使用取决于 ResponseHandler 要不要处理,比如只能写入数据对象存在的对象,默认是忽略存不存在 Class<?> writeClass = null; if (params.getWriteField() != null && StringUtils.hasText(params.getWriteField())) { Field field = ReflectionUtils.findField(gen.getCurrentValue().getClass(), params.getWriteField()); if (field != null) { writeClass = field.getType(); } } // 关闭不存在字段扩展,被写入的字段类型找不到,不扩展 if (!jacksonExpandProperties.isCanExpandToNotExistField() && writeClass == null) { gen.writeObject(bindData); return; } // 翻译为非当前字段,先写入当前字段值再翻译 boolean currField = gen.getOutputContext().getCurrentName().equals(writeField); if (!currField) { gen.writeObject(bindData); gen.writeFieldName(writeField); } if (bindData == null || loadService == null) { gen.writeObject(bindData); return; } // 获取缓存KEY Object[] args = params.getRemoteParams(); int argsLength = args == null ? 0 : args.length; String cacheKey = jacksonExpandProperties.getCachePrefix() + ":" + beanName + ":" + method + ":%s:" + paramsHandler.getCacheKey(bindData, args); Object result = getCacheInfo(cacheKey); if (result != null) { LOG.info("{} Expand cache 命中: {}", beanName, result); gen.writeObject(result); return; } StampedLock lock = lockCache.get(cacheKey, new StampedLock(), Duration.ofSeconds(300)); // 写锁避免同一业务ID重复查询 long stamp = lock.writeLock(); Integer cacheTime = params.getCacheTime(); try { // 多参数组装 Object[] objectParams = new Object[argsLength + 1]; objectParams[0] = paramsHandler.handleVal(bindData); if(objectParams.length > 1){ System.arraycopy(args, 0, objectParams, 1, argsLength); } // 请求翻译结果 Object loadResult = ReflectUtil.invoke(loadService, method, objectParams); if (loadResult != null) { result = this.responseHandler.handle(this.beanName, method, loadResult, writeClass, objectParams); cache.put(String.format(cacheKey, OK), result, Duration.ofSeconds(cacheTime)); } else { LOG.error("【{}】 Expand失败,未找到:{}", beanName, bindData); cache.put(String.format(cacheKey, FAIL), bindData, Duration.ofSeconds(cacheTime)); result = bindData; } } catch (Exception e) { LOG.error("【{}】 Expand异常:", beanName, e); cache.put(String.format(cacheKey, FAIL), bindData, Duration.ofSeconds(cacheTime)); result = bindData; } finally { lock.unlockWrite(stamp); } gen.writeObject(result); } /** * 获取当前字段的path路径 * * @param outputContext * @return */ private String getFieldPath(JsonStreamContext outputContext) { List<String> path = new ArrayList<>(4); while (outputContext != null) { String currentName = outputContext.getCurrentName(); if (StringUtils.hasText(currentName)) { path.add(currentName); } outputContext = outputContext.getParent(); } Collections.reverse(path); return String.join(".", path); } /** * 获取厍信息 * * @param cacheKey 缓存的KEY * @return */ private Object getCacheInfo(String cacheKey) { Object result = cache.get(String.format(cacheKey, OK)); if (result == null) { result = cache.get(String.format(cacheKey, FAIL)); } return result; } @Override public JsonSerializer<?> createContextual(SerializerProvider prov, BeanProperty property) throws JsonMappingException { if (property != null) { Expand load = property.getAnnotation(Expand.class); if (load == null) { throw new RuntimeException("未注解相关 @Expand 注解"); } String bean = load.bean(); Class<? extends ParamsHandler> paramsHandlerClass = load.paramsHandler(); Class<? extends ResponseHandler> responseHandlerClass = load.responseHandler(); String method = load.method(); try { ParamsHandler paramsHandler = paramsHandlerClass.getDeclaredConstructor().newInstance(); ResponseHandler responseHandler = responseHandlerClass.getDeclaredConstructor().newInstance(); int cacheTime = load.cacheTime(); // 额外参数处理 SerializerParam params = paramsHandler.handleAnnotation(property); // 参数处理器没设置,且父注设置了,以父注解为主 if (params.getCacheTime() == null && cacheTime != -1) { params.setCacheTime(cacheTime); } // 缓存时间未设置,取默认 if (params.getCacheTime() == null) { params.
setCacheTime(jacksonExpandProperties.getCacheTimeout());
} if (params.isOpen() == null) { params.setExpand(load.expand()); } return new ExpandSerializer(bean, method, params, paramsHandler, responseHandler); } catch (Exception e) { LOG.error("@Expand error: ", e); } } return prov.findNullValueSerializer(null); } /** * 获取展开参数 * * @param key * @return */ private Set<String> getParam(String key) { RequestAttributes attributes = RequestContextHolder.getRequestAttributes(); if (attributes == null) { return Collections.emptySet(); } ServletRequest request = ((ServletRequestAttributes) attributes).getRequest(); String[] parameterValues = request.getParameterValues(key); if (parameterValues == null) { return Collections.emptySet(); } return Arrays.stream(parameterValues).flatMap(o -> Arrays.stream(o.split(","))) .collect(Collectors.toSet()); } }
src/main/java/com/github/stupdit1t/jackson/expand/serializer/ExpandSerializer.java
stupdit1t-jackson-expand-boot-starter-b492b74
[ { "filename": "src/main/java/com/github/stupdit1t/jackson/expand/handler/params/DefaultParamsHandler.java", "retrieved_chunk": "public class DefaultParamsHandler implements ParamsHandler {\n @Override\n public Object handleVal(Object val) {\n return val;\n }\n @Override\n public SerializerParam handleAnnotation(BeanProperty property) {\n SerializerParam params = new SerializerParam();\n // 用户注解值处理\n Expand expand = property.getAnnotation(Expand.class);", "score": 0.8539618849754333 }, { "filename": "src/main/java/com/github/stupdit1t/jackson/expand/handler/params/ParamsHandler.java", "retrieved_chunk": "package com.github.stupdit1t.jackson.expand.handler.params;\nimport com.fasterxml.jackson.databind.BeanProperty;\nimport com.github.stupdit1t.jackson.expand.domain.SerializerParam;\nimport java.util.StringJoiner;\n/**\n * 参数处理类\n */\npublic interface ParamsHandler {\n /**\n * 处理注解对象的值", "score": 0.8529314994812012 }, { "filename": "src/main/java/com/github/stupdit1t/jackson/expand/handler/params/ParamsHandler.java", "retrieved_chunk": " */\n SerializerParam handleAnnotation(BeanProperty property);\n /**\n * 获取缓存的key\n *\n * @param val 当前值\n * @param annotationVal 注解值\n * @return\n */\n default String getCacheKey(Object val, Object[] annotationVal) {", "score": 0.8426889181137085 }, { "filename": "src/main/java/com/github/stupdit1t/jackson/expand/annotation/Expand.java", "retrieved_chunk": "package com.github.stupdit1t.jackson.expand.annotation;\nimport com.fasterxml.jackson.annotation.JacksonAnnotationsInside;\nimport com.fasterxml.jackson.databind.annotation.JsonSerialize;\nimport com.github.stupdit1t.jackson.expand.handler.params.DefaultParamsHandler;\nimport com.github.stupdit1t.jackson.expand.handler.params.ParamsHandler;\nimport com.github.stupdit1t.jackson.expand.handler.rsp.DefaultResponseHandler;\nimport com.github.stupdit1t.jackson.expand.handler.rsp.ResponseHandler;\nimport com.github.stupdit1t.jackson.expand.serializer.ExpandSerializer;\nimport java.lang.annotation.*;\n/**", "score": 0.8388642072677612 }, { "filename": "src/main/java/com/github/stupdit1t/jackson/expand/cache/ExpandCache.java", "retrieved_chunk": "package com.github.stupdit1t.jackson.expand.cache;\nimport com.github.stupdit1t.jackson.expand.config.JacksonExpandProperties;\nimport com.github.stupdit1t.jackson.expand.util.SpringUtil;\nimport com.github.stupdit1t.jackson.expand.serializer.ExpandSerializer;\nimport org.springframework.util.AntPathMatcher;\nimport java.time.Duration;\nimport java.util.Set;\nimport java.util.StringJoiner;\n/**\n * 缓存抽象", "score": 0.8274906873703003 } ]
java
setCacheTime(jacksonExpandProperties.getCacheTimeout());
package com.github.stupdit1t.jackson.expand.serializer; import com.fasterxml.jackson.core.JsonGenerator; import com.fasterxml.jackson.core.JsonStreamContext; import com.fasterxml.jackson.databind.BeanProperty; import com.fasterxml.jackson.databind.JsonMappingException; import com.fasterxml.jackson.databind.JsonSerializer; import com.fasterxml.jackson.databind.SerializerProvider; import com.fasterxml.jackson.databind.ser.ContextualSerializer; import com.github.stupdit1t.jackson.expand.annotation.Expand; import com.github.stupdit1t.jackson.expand.cache.ExpandCache; import com.github.stupdit1t.jackson.expand.cache.LocalCache; import com.github.stupdit1t.jackson.expand.config.JacksonExpandProperties; import com.github.stupdit1t.jackson.expand.domain.ExpandStrategy; import com.github.stupdit1t.jackson.expand.domain.SerializerParam; import com.github.stupdit1t.jackson.expand.handler.params.ParamsHandler; import com.github.stupdit1t.jackson.expand.handler.rsp.ResponseHandler; import com.github.stupdit1t.jackson.expand.util.ReflectUtil; import com.github.stupdit1t.jackson.expand.util.SpringUtil; import org.slf4j.Logger; import org.slf4j.LoggerFactory; import org.springframework.util.ReflectionUtils; import org.springframework.util.StringUtils; import org.springframework.web.context.request.RequestAttributes; import org.springframework.web.context.request.RequestContextHolder; import org.springframework.web.context.request.ServletRequestAttributes; import javax.servlet.ServletRequest; import java.io.IOException; import java.lang.reflect.Field; import java.time.Duration; import java.util.*; import java.util.concurrent.locks.StampedLock; import java.util.stream.Collectors; public class ExpandSerializer extends JsonSerializer<Object> implements ContextualSerializer { private static final Logger LOG = LoggerFactory.getLogger(ExpandSerializer.class); /** * 成功数据 */ public static final String OK = "OK"; /** * 失败数据 */ public static final String FAIL = "FAIL"; /** * 缓存 */ private static ExpandCache cache; /** * 配置 */ private static JacksonExpandProperties jacksonExpandProperties; /** * 本地锁缓存,防止同时查询 */ private static final LocalCache lockCache = new LocalCache(); /** * 远程调用服务 */ private Object loadService; /** * 方法 */ private String method; /** * 注解参数处理 */ private SerializerParam params; /** * 返回结果处理类 */ private ParamsHandler paramsHandler; /** * 返回结果处理类 */ private ResponseHandler responseHandler; /** * bean名称 */ private String beanName; public ExpandSerializer() { super(); if (cache == null) { synchronized (ExpandSerializer.class) { if (cache == null) { cache = SpringUtil.getBean(ExpandCache.class); jacksonExpandProperties = SpringUtil.getBean(JacksonExpandProperties.class); } } } } public ExpandSerializer(String beanName, String method, SerializerParam params, ParamsHandler paramsHandler, ResponseHandler otherResponseHandler) { this(); this.loadService = SpringUtil.getBean(beanName); this.method = method; this.params = params; this.responseHandler = otherResponseHandler; this.paramsHandler = paramsHandler; this.beanName = beanName; } @Override public void serialize(Object bindData, JsonGenerator gen, SerializerProvider serializers) throws IOException { String writeFieldPath = getFieldPath(gen.getOutputContext()); // 统一path替换 String dynamicExpandCommonPrefix = jacksonExpandProperties.getDynamicExpandCommonPrefix(); if (StringUtils.hasText(dynamicExpandCommonPrefix) && writeFieldPath.startsWith(dynamicExpandCommonPrefix)) { writeFieldPath = writeFieldPath.substring(dynamicExpandCommonPrefix.length() + 1); } // 是否展开 boolean expand; // 动态展开开启,判断是否展开 boolean dynamicExpand = jacksonExpandProperties.isDynamicExpand(); if (dynamicExpand) { Set<String> needExpandField = getParam(jacksonExpandProperties.getDynamicExpandParameterName()); // 如果代码里设置不展开,动态展开也不生效 expand = needExpandField.contains(writeFieldPath) && params.isOpen(); } else { expand = params.isOpen(); } if (!expand) { gen.writeObject(bindData); return; } // 判断要写入的字段 String writeField = gen.getOutputContext().getCurrentName(); if (jacksonExpandProperties.getExpandStrategy() == ExpandStrategy.COVER) { writeField = gen.getOutputContext().getCurrentName(); } else if (jacksonExpandProperties.getExpandStrategy() == ExpandStrategy.COPY) { writeField = String.format(jacksonExpandProperties.getCopyStrategyFormat(), gen.getOutputContext().getCurrentName()); } // 自定义要写入的优先级最高 if (StringUtils.hasText(params.getWriteField())) { writeField = params.getWriteField(); } // 设置理论上的响应类型,要不要使用取决于 ResponseHandler 要不要处理,比如只能写入数据对象存在的对象,默认是忽略存不存在 Class<?> writeClass = null; if (params.getWriteField() != null && StringUtils.hasText(params.getWriteField())) { Field field = ReflectionUtils.findField(gen.getCurrentValue().getClass(), params.getWriteField()); if (field != null) { writeClass = field.getType(); } } // 关闭不存在字段扩展,被写入的字段类型找不到,不扩展 if (!jacksonExpandProperties.isCanExpandToNotExistField() && writeClass == null) { gen.writeObject(bindData); return; } // 翻译为非当前字段,先写入当前字段值再翻译 boolean currField = gen.getOutputContext().getCurrentName().equals(writeField); if (!currField) { gen.writeObject(bindData); gen.writeFieldName(writeField); } if (bindData == null || loadService == null) { gen.writeObject(bindData); return; } // 获取缓存KEY Object[] args = params.getRemoteParams(); int argsLength = args == null ? 0 : args.length; String cacheKey = jacksonExpandProperties.getCachePrefix() + ":" + beanName + ":" + method + ":%s:" + paramsHandler.getCacheKey(bindData, args); Object result = getCacheInfo(cacheKey); if (result != null) { LOG.info("{} Expand cache 命中: {}", beanName, result); gen.writeObject(result); return; } StampedLock lock = lockCache.get(cacheKey, new StampedLock(), Duration.ofSeconds(300)); // 写锁避免同一业务ID重复查询 long stamp = lock.writeLock(); Integer cacheTime = params.getCacheTime(); try { // 多参数组装 Object[] objectParams = new Object[argsLength + 1]; objectParams[0] = paramsHandler.handleVal(bindData); if(objectParams.length > 1){ System.arraycopy(args, 0, objectParams, 1, argsLength); } // 请求翻译结果 Object loadResult = ReflectUtil.invoke(loadService, method, objectParams); if (loadResult != null) { result = this.responseHandler.handle(this.beanName, method, loadResult, writeClass, objectParams);
cache.put(String.format(cacheKey, OK), result, Duration.ofSeconds(cacheTime));
} else { LOG.error("【{}】 Expand失败,未找到:{}", beanName, bindData); cache.put(String.format(cacheKey, FAIL), bindData, Duration.ofSeconds(cacheTime)); result = bindData; } } catch (Exception e) { LOG.error("【{}】 Expand异常:", beanName, e); cache.put(String.format(cacheKey, FAIL), bindData, Duration.ofSeconds(cacheTime)); result = bindData; } finally { lock.unlockWrite(stamp); } gen.writeObject(result); } /** * 获取当前字段的path路径 * * @param outputContext * @return */ private String getFieldPath(JsonStreamContext outputContext) { List<String> path = new ArrayList<>(4); while (outputContext != null) { String currentName = outputContext.getCurrentName(); if (StringUtils.hasText(currentName)) { path.add(currentName); } outputContext = outputContext.getParent(); } Collections.reverse(path); return String.join(".", path); } /** * 获取厍信息 * * @param cacheKey 缓存的KEY * @return */ private Object getCacheInfo(String cacheKey) { Object result = cache.get(String.format(cacheKey, OK)); if (result == null) { result = cache.get(String.format(cacheKey, FAIL)); } return result; } @Override public JsonSerializer<?> createContextual(SerializerProvider prov, BeanProperty property) throws JsonMappingException { if (property != null) { Expand load = property.getAnnotation(Expand.class); if (load == null) { throw new RuntimeException("未注解相关 @Expand 注解"); } String bean = load.bean(); Class<? extends ParamsHandler> paramsHandlerClass = load.paramsHandler(); Class<? extends ResponseHandler> responseHandlerClass = load.responseHandler(); String method = load.method(); try { ParamsHandler paramsHandler = paramsHandlerClass.getDeclaredConstructor().newInstance(); ResponseHandler responseHandler = responseHandlerClass.getDeclaredConstructor().newInstance(); int cacheTime = load.cacheTime(); // 额外参数处理 SerializerParam params = paramsHandler.handleAnnotation(property); // 参数处理器没设置,且父注设置了,以父注解为主 if (params.getCacheTime() == null && cacheTime != -1) { params.setCacheTime(cacheTime); } // 缓存时间未设置,取默认 if (params.getCacheTime() == null) { params.setCacheTime(jacksonExpandProperties.getCacheTimeout()); } if (params.isOpen() == null) { params.setExpand(load.expand()); } return new ExpandSerializer(bean, method, params, paramsHandler, responseHandler); } catch (Exception e) { LOG.error("@Expand error: ", e); } } return prov.findNullValueSerializer(null); } /** * 获取展开参数 * * @param key * @return */ private Set<String> getParam(String key) { RequestAttributes attributes = RequestContextHolder.getRequestAttributes(); if (attributes == null) { return Collections.emptySet(); } ServletRequest request = ((ServletRequestAttributes) attributes).getRequest(); String[] parameterValues = request.getParameterValues(key); if (parameterValues == null) { return Collections.emptySet(); } return Arrays.stream(parameterValues).flatMap(o -> Arrays.stream(o.split(","))) .collect(Collectors.toSet()); } }
src/main/java/com/github/stupdit1t/jackson/expand/serializer/ExpandSerializer.java
stupdit1t-jackson-expand-boot-starter-b492b74
[ { "filename": "src/main/java/com/github/stupdit1t/jackson/expand/handler/rsp/DefaultResponseHandler.java", "retrieved_chunk": " * @param rsp 当前返回值\n * @param toClass 要填充字段的类型\n * @param params 当前方法参数\n * @return\n */\n @Override\n public Object handle(String bean, String method, Object rsp, Class<?> toClass, Object... params) {\n return rsp;\n }\n}", "score": 0.8255398273468018 }, { "filename": "src/main/java/com/github/stupdit1t/jackson/expand/handler/rsp/ResponseHandler.java", "retrieved_chunk": "package com.github.stupdit1t.jackson.expand.handler.rsp;\n/**\n * 返回处理第三方响应\n */\npublic interface ResponseHandler {\n Object handle(String bean, String method, Object rsp, Class<?> writeClass, Object... params);\n}", "score": 0.8126899003982544 }, { "filename": "src/main/java/com/github/stupdit1t/jackson/expand/util/ReflectUtil.java", "retrieved_chunk": " *\n * @param obj\n * @param methodName\n * @param args\n * @return\n */\n public static Object invoke(Object obj, String methodName, Object[] args) throws InvocationTargetException, IllegalAccessException {\n String cacheKey = obj.getClass().getName() + methodName;\n final Method method = METHODS_CACHE.computeIfAbsent(cacheKey, (key) -> getMethod(obj.getClass(), methodName, args));\n if (null == method) {", "score": 0.8034810423851013 }, { "filename": "src/main/java/com/github/stupdit1t/jackson/expand/util/ReflectUtil.java", "retrieved_chunk": " WRAPPER_PRIMITIVE_MAP.put(Integer.class, int.class);\n WRAPPER_PRIMITIVE_MAP.put(Long.class, long.class);\n WRAPPER_PRIMITIVE_MAP.put(Short.class, short.class);\n }\n /**\n * 反射方法缓存\n */\n private static final Map<String, Method> METHODS_CACHE = new ConcurrentHashMap<>();\n /**\n * 反射调用", "score": 0.8010255098342896 }, { "filename": "src/main/java/com/github/stupdit1t/jackson/expand/cache/ExpandCache.java", "retrieved_chunk": " key.add(String.valueOf(bindData));\n for (Object subVal : annotationVal) {\n key.add(String.valueOf(subVal));\n }\n String cacheKey = properties.getCachePrefix() + \":\" + beanName + \":\" + method + \":%s:\" + key.toString();\n delete(String.format(cacheKey, ExpandSerializer.OK));\n delete(String.format(cacheKey, ExpandSerializer.FAIL));\n }\n /**\n * 模糊匹配key", "score": 0.7994896769523621 } ]
java
cache.put(String.format(cacheKey, OK), result, Duration.ofSeconds(cacheTime));
package com.github.stupdit1t.jackson.expand.serializer; import com.fasterxml.jackson.core.JsonGenerator; import com.fasterxml.jackson.core.JsonStreamContext; import com.fasterxml.jackson.databind.BeanProperty; import com.fasterxml.jackson.databind.JsonMappingException; import com.fasterxml.jackson.databind.JsonSerializer; import com.fasterxml.jackson.databind.SerializerProvider; import com.fasterxml.jackson.databind.ser.ContextualSerializer; import com.github.stupdit1t.jackson.expand.annotation.Expand; import com.github.stupdit1t.jackson.expand.cache.ExpandCache; import com.github.stupdit1t.jackson.expand.cache.LocalCache; import com.github.stupdit1t.jackson.expand.config.JacksonExpandProperties; import com.github.stupdit1t.jackson.expand.domain.ExpandStrategy; import com.github.stupdit1t.jackson.expand.domain.SerializerParam; import com.github.stupdit1t.jackson.expand.handler.params.ParamsHandler; import com.github.stupdit1t.jackson.expand.handler.rsp.ResponseHandler; import com.github.stupdit1t.jackson.expand.util.ReflectUtil; import com.github.stupdit1t.jackson.expand.util.SpringUtil; import org.slf4j.Logger; import org.slf4j.LoggerFactory; import org.springframework.util.ReflectionUtils; import org.springframework.util.StringUtils; import org.springframework.web.context.request.RequestAttributes; import org.springframework.web.context.request.RequestContextHolder; import org.springframework.web.context.request.ServletRequestAttributes; import javax.servlet.ServletRequest; import java.io.IOException; import java.lang.reflect.Field; import java.time.Duration; import java.util.*; import java.util.concurrent.locks.StampedLock; import java.util.stream.Collectors; public class ExpandSerializer extends JsonSerializer<Object> implements ContextualSerializer { private static final Logger LOG = LoggerFactory.getLogger(ExpandSerializer.class); /** * 成功数据 */ public static final String OK = "OK"; /** * 失败数据 */ public static final String FAIL = "FAIL"; /** * 缓存 */ private static ExpandCache cache; /** * 配置 */ private static JacksonExpandProperties jacksonExpandProperties; /** * 本地锁缓存,防止同时查询 */ private static final LocalCache lockCache = new LocalCache(); /** * 远程调用服务 */ private Object loadService; /** * 方法 */ private String method; /** * 注解参数处理 */ private SerializerParam params; /** * 返回结果处理类 */ private ParamsHandler paramsHandler; /** * 返回结果处理类 */ private ResponseHandler responseHandler; /** * bean名称 */ private String beanName; public ExpandSerializer() { super(); if (cache == null) { synchronized (ExpandSerializer.class) { if (cache == null) { cache = SpringUtil.getBean(ExpandCache.class); jacksonExpandProperties = SpringUtil.getBean(JacksonExpandProperties.class); } } } } public ExpandSerializer(String beanName, String method, SerializerParam params, ParamsHandler paramsHandler, ResponseHandler otherResponseHandler) { this(); this.loadService = SpringUtil.getBean(beanName); this.method = method; this.params = params; this.responseHandler = otherResponseHandler; this.paramsHandler = paramsHandler; this.beanName = beanName; } @Override public void serialize(Object bindData, JsonGenerator gen, SerializerProvider serializers) throws IOException { String writeFieldPath = getFieldPath(gen.getOutputContext()); // 统一path替换 String dynamicExpandCommonPrefix = jacksonExpandProperties.getDynamicExpandCommonPrefix(); if (StringUtils.hasText(dynamicExpandCommonPrefix) && writeFieldPath.startsWith(dynamicExpandCommonPrefix)) { writeFieldPath = writeFieldPath.substring(dynamicExpandCommonPrefix.length() + 1); } // 是否展开 boolean expand; // 动态展开开启,判断是否展开 boolean dynamicExpand = jacksonExpandProperties.isDynamicExpand(); if (dynamicExpand) { Set<String> needExpandField = getParam(jacksonExpandProperties.getDynamicExpandParameterName()); // 如果代码里设置不展开,动态展开也不生效 expand = needExpandField.contains(writeFieldPath) && params.isOpen(); } else { expand = params.isOpen(); } if (!expand) { gen.writeObject(bindData); return; } // 判断要写入的字段 String writeField = gen.getOutputContext().getCurrentName(); if (
jacksonExpandProperties.getExpandStrategy() == ExpandStrategy.COVER) {
writeField = gen.getOutputContext().getCurrentName(); } else if (jacksonExpandProperties.getExpandStrategy() == ExpandStrategy.COPY) { writeField = String.format(jacksonExpandProperties.getCopyStrategyFormat(), gen.getOutputContext().getCurrentName()); } // 自定义要写入的优先级最高 if (StringUtils.hasText(params.getWriteField())) { writeField = params.getWriteField(); } // 设置理论上的响应类型,要不要使用取决于 ResponseHandler 要不要处理,比如只能写入数据对象存在的对象,默认是忽略存不存在 Class<?> writeClass = null; if (params.getWriteField() != null && StringUtils.hasText(params.getWriteField())) { Field field = ReflectionUtils.findField(gen.getCurrentValue().getClass(), params.getWriteField()); if (field != null) { writeClass = field.getType(); } } // 关闭不存在字段扩展,被写入的字段类型找不到,不扩展 if (!jacksonExpandProperties.isCanExpandToNotExistField() && writeClass == null) { gen.writeObject(bindData); return; } // 翻译为非当前字段,先写入当前字段值再翻译 boolean currField = gen.getOutputContext().getCurrentName().equals(writeField); if (!currField) { gen.writeObject(bindData); gen.writeFieldName(writeField); } if (bindData == null || loadService == null) { gen.writeObject(bindData); return; } // 获取缓存KEY Object[] args = params.getRemoteParams(); int argsLength = args == null ? 0 : args.length; String cacheKey = jacksonExpandProperties.getCachePrefix() + ":" + beanName + ":" + method + ":%s:" + paramsHandler.getCacheKey(bindData, args); Object result = getCacheInfo(cacheKey); if (result != null) { LOG.info("{} Expand cache 命中: {}", beanName, result); gen.writeObject(result); return; } StampedLock lock = lockCache.get(cacheKey, new StampedLock(), Duration.ofSeconds(300)); // 写锁避免同一业务ID重复查询 long stamp = lock.writeLock(); Integer cacheTime = params.getCacheTime(); try { // 多参数组装 Object[] objectParams = new Object[argsLength + 1]; objectParams[0] = paramsHandler.handleVal(bindData); if(objectParams.length > 1){ System.arraycopy(args, 0, objectParams, 1, argsLength); } // 请求翻译结果 Object loadResult = ReflectUtil.invoke(loadService, method, objectParams); if (loadResult != null) { result = this.responseHandler.handle(this.beanName, method, loadResult, writeClass, objectParams); cache.put(String.format(cacheKey, OK), result, Duration.ofSeconds(cacheTime)); } else { LOG.error("【{}】 Expand失败,未找到:{}", beanName, bindData); cache.put(String.format(cacheKey, FAIL), bindData, Duration.ofSeconds(cacheTime)); result = bindData; } } catch (Exception e) { LOG.error("【{}】 Expand异常:", beanName, e); cache.put(String.format(cacheKey, FAIL), bindData, Duration.ofSeconds(cacheTime)); result = bindData; } finally { lock.unlockWrite(stamp); } gen.writeObject(result); } /** * 获取当前字段的path路径 * * @param outputContext * @return */ private String getFieldPath(JsonStreamContext outputContext) { List<String> path = new ArrayList<>(4); while (outputContext != null) { String currentName = outputContext.getCurrentName(); if (StringUtils.hasText(currentName)) { path.add(currentName); } outputContext = outputContext.getParent(); } Collections.reverse(path); return String.join(".", path); } /** * 获取厍信息 * * @param cacheKey 缓存的KEY * @return */ private Object getCacheInfo(String cacheKey) { Object result = cache.get(String.format(cacheKey, OK)); if (result == null) { result = cache.get(String.format(cacheKey, FAIL)); } return result; } @Override public JsonSerializer<?> createContextual(SerializerProvider prov, BeanProperty property) throws JsonMappingException { if (property != null) { Expand load = property.getAnnotation(Expand.class); if (load == null) { throw new RuntimeException("未注解相关 @Expand 注解"); } String bean = load.bean(); Class<? extends ParamsHandler> paramsHandlerClass = load.paramsHandler(); Class<? extends ResponseHandler> responseHandlerClass = load.responseHandler(); String method = load.method(); try { ParamsHandler paramsHandler = paramsHandlerClass.getDeclaredConstructor().newInstance(); ResponseHandler responseHandler = responseHandlerClass.getDeclaredConstructor().newInstance(); int cacheTime = load.cacheTime(); // 额外参数处理 SerializerParam params = paramsHandler.handleAnnotation(property); // 参数处理器没设置,且父注设置了,以父注解为主 if (params.getCacheTime() == null && cacheTime != -1) { params.setCacheTime(cacheTime); } // 缓存时间未设置,取默认 if (params.getCacheTime() == null) { params.setCacheTime(jacksonExpandProperties.getCacheTimeout()); } if (params.isOpen() == null) { params.setExpand(load.expand()); } return new ExpandSerializer(bean, method, params, paramsHandler, responseHandler); } catch (Exception e) { LOG.error("@Expand error: ", e); } } return prov.findNullValueSerializer(null); } /** * 获取展开参数 * * @param key * @return */ private Set<String> getParam(String key) { RequestAttributes attributes = RequestContextHolder.getRequestAttributes(); if (attributes == null) { return Collections.emptySet(); } ServletRequest request = ((ServletRequestAttributes) attributes).getRequest(); String[] parameterValues = request.getParameterValues(key); if (parameterValues == null) { return Collections.emptySet(); } return Arrays.stream(parameterValues).flatMap(o -> Arrays.stream(o.split(","))) .collect(Collectors.toSet()); } }
src/main/java/com/github/stupdit1t/jackson/expand/serializer/ExpandSerializer.java
stupdit1t-jackson-expand-boot-starter-b492b74
[ { "filename": "src/main/java/com/github/stupdit1t/jackson/expand/handler/params/DefaultParamsHandler.java", "retrieved_chunk": " if (expand != null) {\n if (StringUtils.hasText(expand.to())) {\n params.setWriteField(expand.to());\n }\n params.setExpand(expand.expand());\n }\n return params;\n }\n}", "score": 0.8674397468566895 }, { "filename": "src/main/java/com/github/stupdit1t/jackson/expand/domain/ExpandStrategy.java", "retrieved_chunk": "package com.github.stupdit1t.jackson.expand.domain;\n/**\n * 展开策略\n */\npublic enum ExpandStrategy {\n /**\n * 覆盖当前字段\n */\n COVER,\n /**", "score": 0.8352632522583008 }, { "filename": "src/main/java/com/github/stupdit1t/jackson/expand/domain/SerializerParam.java", "retrieved_chunk": " * 写入字段\n */\n private String writeField;\n /**\n * 值缓存时间\n */\n private Integer cacheTime;\n /**\n * 是否展开\n */", "score": 0.8301984667778015 }, { "filename": "src/main/java/com/github/stupdit1t/jackson/expand/domain/SerializerParam.java", "retrieved_chunk": " private Boolean expand;\n public Object[] getRemoteParams() {\n return remoteParams;\n }\n public void setRemoteParams(Object[] remoteParams) {\n this.remoteParams = remoteParams;\n }\n public String getWriteField() {\n return writeField;\n }", "score": 0.8266885876655579 }, { "filename": "src/main/java/com/github/stupdit1t/jackson/expand/config/JacksonExpandProperties.java", "retrieved_chunk": " */\n private String dynamicExpandParameterName = \"expand\";\n /**\n * 动态展开 统一数据的Path前缀,比如前缀是 data.body. 如果配置 expand=userId, 相当于是expnad=data.body.userId, 默认无\n */\n private String dynamicExpandCommonPrefix;\n /**\n * 展开策略, 默认覆盖\n */\n private ExpandStrategy expandStrategy = ExpandStrategy.COVER;", "score": 0.8239043354988098 } ]
java
jacksonExpandProperties.getExpandStrategy() == ExpandStrategy.COVER) {
package com.github.stupdit1t.jackson.expand.serializer; import com.fasterxml.jackson.core.JsonGenerator; import com.fasterxml.jackson.core.JsonStreamContext; import com.fasterxml.jackson.databind.BeanProperty; import com.fasterxml.jackson.databind.JsonMappingException; import com.fasterxml.jackson.databind.JsonSerializer; import com.fasterxml.jackson.databind.SerializerProvider; import com.fasterxml.jackson.databind.ser.ContextualSerializer; import com.github.stupdit1t.jackson.expand.annotation.Expand; import com.github.stupdit1t.jackson.expand.cache.ExpandCache; import com.github.stupdit1t.jackson.expand.cache.LocalCache; import com.github.stupdit1t.jackson.expand.config.JacksonExpandProperties; import com.github.stupdit1t.jackson.expand.domain.ExpandStrategy; import com.github.stupdit1t.jackson.expand.domain.SerializerParam; import com.github.stupdit1t.jackson.expand.handler.params.ParamsHandler; import com.github.stupdit1t.jackson.expand.handler.rsp.ResponseHandler; import com.github.stupdit1t.jackson.expand.util.ReflectUtil; import com.github.stupdit1t.jackson.expand.util.SpringUtil; import org.slf4j.Logger; import org.slf4j.LoggerFactory; import org.springframework.util.ReflectionUtils; import org.springframework.util.StringUtils; import org.springframework.web.context.request.RequestAttributes; import org.springframework.web.context.request.RequestContextHolder; import org.springframework.web.context.request.ServletRequestAttributes; import javax.servlet.ServletRequest; import java.io.IOException; import java.lang.reflect.Field; import java.time.Duration; import java.util.*; import java.util.concurrent.locks.StampedLock; import java.util.stream.Collectors; public class ExpandSerializer extends JsonSerializer<Object> implements ContextualSerializer { private static final Logger LOG = LoggerFactory.getLogger(ExpandSerializer.class); /** * 成功数据 */ public static final String OK = "OK"; /** * 失败数据 */ public static final String FAIL = "FAIL"; /** * 缓存 */ private static ExpandCache cache; /** * 配置 */ private static JacksonExpandProperties jacksonExpandProperties; /** * 本地锁缓存,防止同时查询 */ private static final LocalCache lockCache = new LocalCache(); /** * 远程调用服务 */ private Object loadService; /** * 方法 */ private String method; /** * 注解参数处理 */ private SerializerParam params; /** * 返回结果处理类 */ private ParamsHandler paramsHandler; /** * 返回结果处理类 */ private ResponseHandler responseHandler; /** * bean名称 */ private String beanName; public ExpandSerializer() { super(); if (cache == null) { synchronized (ExpandSerializer.class) { if (cache == null) { cache = SpringUtil.getBean(ExpandCache.class); jacksonExpandProperties = SpringUtil.getBean(JacksonExpandProperties.class); } } } } public ExpandSerializer(String beanName, String method, SerializerParam params, ParamsHandler paramsHandler, ResponseHandler otherResponseHandler) { this(); this.loadService = SpringUtil.getBean(beanName); this.method = method; this.params = params; this.responseHandler = otherResponseHandler; this.paramsHandler = paramsHandler; this.beanName = beanName; } @Override public void serialize(Object bindData, JsonGenerator gen, SerializerProvider serializers) throws IOException { String writeFieldPath = getFieldPath(gen.getOutputContext()); // 统一path替换 String dynamicExpandCommonPrefix = jacksonExpandProperties.getDynamicExpandCommonPrefix(); if (StringUtils.hasText(dynamicExpandCommonPrefix) && writeFieldPath.startsWith(dynamicExpandCommonPrefix)) { writeFieldPath = writeFieldPath.substring(dynamicExpandCommonPrefix.length() + 1); } // 是否展开 boolean expand; // 动态展开开启,判断是否展开 boolean dynamicExpand = jacksonExpandProperties.isDynamicExpand(); if (dynamicExpand) { Set<String> needExpandField = getParam(jacksonExpandProperties.getDynamicExpandParameterName()); // 如果代码里设置不展开,动态展开也不生效 expand = needExpandField.contains(writeFieldPath) && params.isOpen(); } else { expand = params.isOpen(); } if (!expand) { gen.writeObject(bindData); return; } // 判断要写入的字段 String writeField = gen.getOutputContext().getCurrentName(); if (jacksonExpandProperties.getExpandStrategy() == ExpandStrategy.COVER) { writeField = gen.getOutputContext().getCurrentName(); } else if (jacksonExpandProperties.getExpandStrategy() == ExpandStrategy.COPY) { writeField = String.format(jacksonExpandProperties.getCopyStrategyFormat(), gen.getOutputContext().getCurrentName()); } // 自定义要写入的优先级最高 if (StringUtils.hasText(params.getWriteField())) { writeField = params.getWriteField(); } // 设置理论上的响应类型,要不要使用取决于 ResponseHandler 要不要处理,比如只能写入数据对象存在的对象,默认是忽略存不存在 Class<?> writeClass = null; if (params.getWriteField() != null && StringUtils.hasText(params.getWriteField())) { Field field = ReflectionUtils.findField(gen.getCurrentValue().getClass(), params.getWriteField()); if (field != null) { writeClass = field.getType(); } } // 关闭不存在字段扩展,被写入的字段类型找不到,不扩展 if (!jacksonExpandProperties.isCanExpandToNotExistField() && writeClass == null) { gen.writeObject(bindData); return; } // 翻译为非当前字段,先写入当前字段值再翻译 boolean currField = gen.getOutputContext().getCurrentName().equals(writeField); if (!currField) { gen.writeObject(bindData); gen.writeFieldName(writeField); } if (bindData == null || loadService == null) { gen.writeObject(bindData); return; } // 获取缓存KEY Object[] args = params.getRemoteParams(); int argsLength = args == null ? 0 : args.length; String cacheKey = jacksonExpandProperties.getCachePrefix() + ":" + beanName + ":" + method + ":%s:" + paramsHandler.getCacheKey(bindData, args); Object result = getCacheInfo(cacheKey); if (result != null) { LOG.info("{} Expand cache 命中: {}", beanName, result); gen.writeObject(result); return; } StampedLock lock = lockCache.get(cacheKey, new StampedLock(), Duration.ofSeconds(300)); // 写锁避免同一业务ID重复查询 long stamp = lock.writeLock(); Integer cacheTime = params.getCacheTime(); try { // 多参数组装 Object[] objectParams = new Object[argsLength + 1]; objectParams[0] = paramsHandler.handleVal(bindData); if(objectParams.length > 1){ System.arraycopy(args, 0, objectParams, 1, argsLength); } // 请求翻译结果 Object loadResult = ReflectUtil.invoke(loadService, method, objectParams); if (loadResult != null) { result = this.responseHandler.handle(this.beanName, method, loadResult, writeClass, objectParams); cache.put(String.format(cacheKey, OK), result, Duration.ofSeconds(cacheTime)); } else { LOG.error("【{}】 Expand失败,未找到:{}", beanName, bindData); cache.put(String.format(cacheKey, FAIL), bindData, Duration.ofSeconds(cacheTime)); result = bindData; } } catch (Exception e) { LOG.error("【{}】 Expand异常:", beanName, e); cache.put(String.format(cacheKey, FAIL), bindData, Duration.ofSeconds(cacheTime)); result = bindData; } finally { lock.unlockWrite(stamp); } gen.writeObject(result); } /** * 获取当前字段的path路径 * * @param outputContext * @return */ private String getFieldPath(JsonStreamContext outputContext) { List<String> path = new ArrayList<>(4); while (outputContext != null) { String currentName = outputContext.getCurrentName(); if (StringUtils.hasText(currentName)) { path.add(currentName); } outputContext = outputContext.getParent(); } Collections.reverse(path); return String.join(".", path); } /** * 获取厍信息 * * @param cacheKey 缓存的KEY * @return */ private Object getCacheInfo(String cacheKey) { Object result =
cache.get(String.format(cacheKey, OK));
if (result == null) { result = cache.get(String.format(cacheKey, FAIL)); } return result; } @Override public JsonSerializer<?> createContextual(SerializerProvider prov, BeanProperty property) throws JsonMappingException { if (property != null) { Expand load = property.getAnnotation(Expand.class); if (load == null) { throw new RuntimeException("未注解相关 @Expand 注解"); } String bean = load.bean(); Class<? extends ParamsHandler> paramsHandlerClass = load.paramsHandler(); Class<? extends ResponseHandler> responseHandlerClass = load.responseHandler(); String method = load.method(); try { ParamsHandler paramsHandler = paramsHandlerClass.getDeclaredConstructor().newInstance(); ResponseHandler responseHandler = responseHandlerClass.getDeclaredConstructor().newInstance(); int cacheTime = load.cacheTime(); // 额外参数处理 SerializerParam params = paramsHandler.handleAnnotation(property); // 参数处理器没设置,且父注设置了,以父注解为主 if (params.getCacheTime() == null && cacheTime != -1) { params.setCacheTime(cacheTime); } // 缓存时间未设置,取默认 if (params.getCacheTime() == null) { params.setCacheTime(jacksonExpandProperties.getCacheTimeout()); } if (params.isOpen() == null) { params.setExpand(load.expand()); } return new ExpandSerializer(bean, method, params, paramsHandler, responseHandler); } catch (Exception e) { LOG.error("@Expand error: ", e); } } return prov.findNullValueSerializer(null); } /** * 获取展开参数 * * @param key * @return */ private Set<String> getParam(String key) { RequestAttributes attributes = RequestContextHolder.getRequestAttributes(); if (attributes == null) { return Collections.emptySet(); } ServletRequest request = ((ServletRequestAttributes) attributes).getRequest(); String[] parameterValues = request.getParameterValues(key); if (parameterValues == null) { return Collections.emptySet(); } return Arrays.stream(parameterValues).flatMap(o -> Arrays.stream(o.split(","))) .collect(Collectors.toSet()); } }
src/main/java/com/github/stupdit1t/jackson/expand/serializer/ExpandSerializer.java
stupdit1t-jackson-expand-boot-starter-b492b74
[ { "filename": "src/main/java/com/github/stupdit1t/jackson/expand/config/JacksonExpandProperties.java", "retrieved_chunk": " * 缓存Key前缀\n */\n private String cachePrefix = \"Expand\";\n /**\n * 缓存时间\n * <p>\n * 单位秒\n */\n private Integer cacheTimeout = 300;\n /**", "score": 0.7973734140396118 }, { "filename": "src/main/java/com/github/stupdit1t/jackson/expand/cache/LocalCache.java", "retrieved_chunk": " }\n public <T> T get(String cacheKey, T value, Duration timeout) {\n T val = (T) cacheMap.computeIfAbsent(cacheKey, (key) -> {\n scheduleExpiration(cacheKey, timeout.toMillis());\n return value;\n });\n return val;\n }\n @Override\n public Set<String> keys(String pattern) {", "score": 0.7932621240615845 }, { "filename": "src/main/java/com/github/stupdit1t/jackson/expand/cache/ExpandCache.java", "retrieved_chunk": " key.add(String.valueOf(bindData));\n for (Object subVal : annotationVal) {\n key.add(String.valueOf(subVal));\n }\n String cacheKey = properties.getCachePrefix() + \":\" + beanName + \":\" + method + \":%s:\" + key.toString();\n delete(String.format(cacheKey, ExpandSerializer.OK));\n delete(String.format(cacheKey, ExpandSerializer.FAIL));\n }\n /**\n * 模糊匹配key", "score": 0.7788949012756348 }, { "filename": "src/main/java/com/github/stupdit1t/jackson/expand/cache/ExpandCache.java", "retrieved_chunk": " /**\n * 获取缓存\n *\n * @param key\n * @return\n */\n <T> T get(String key);\n /**\n * 列出匹配的的key\n *", "score": 0.7747541666030884 }, { "filename": "src/main/java/com/github/stupdit1t/jackson/expand/cache/LocalCache.java", "retrieved_chunk": " * 简单的本地缓存实现\n */\npublic class LocalCache implements ExpandCache {\n private static final Logger LOG = LoggerFactory.getLogger(ExpandSerializer.class);\n private static final Timer timer = new Timer();\n private final Map<String, Object> cacheMap;\n public LocalCache() {\n this.cacheMap = new ConcurrentHashMap<>();\n }\n public LocalCache(Map<String, Object> cacheMap) {", "score": 0.7725856304168701 } ]
java
cache.get(String.format(cacheKey, OK));
package com.github.stupdit1t.jackson.expand.serializer; import com.fasterxml.jackson.core.JsonGenerator; import com.fasterxml.jackson.core.JsonStreamContext; import com.fasterxml.jackson.databind.BeanProperty; import com.fasterxml.jackson.databind.JsonMappingException; import com.fasterxml.jackson.databind.JsonSerializer; import com.fasterxml.jackson.databind.SerializerProvider; import com.fasterxml.jackson.databind.ser.ContextualSerializer; import com.github.stupdit1t.jackson.expand.annotation.Expand; import com.github.stupdit1t.jackson.expand.cache.ExpandCache; import com.github.stupdit1t.jackson.expand.cache.LocalCache; import com.github.stupdit1t.jackson.expand.config.JacksonExpandProperties; import com.github.stupdit1t.jackson.expand.domain.ExpandStrategy; import com.github.stupdit1t.jackson.expand.domain.SerializerParam; import com.github.stupdit1t.jackson.expand.handler.params.ParamsHandler; import com.github.stupdit1t.jackson.expand.handler.rsp.ResponseHandler; import com.github.stupdit1t.jackson.expand.util.ReflectUtil; import com.github.stupdit1t.jackson.expand.util.SpringUtil; import org.slf4j.Logger; import org.slf4j.LoggerFactory; import org.springframework.util.ReflectionUtils; import org.springframework.util.StringUtils; import org.springframework.web.context.request.RequestAttributes; import org.springframework.web.context.request.RequestContextHolder; import org.springframework.web.context.request.ServletRequestAttributes; import javax.servlet.ServletRequest; import java.io.IOException; import java.lang.reflect.Field; import java.time.Duration; import java.util.*; import java.util.concurrent.locks.StampedLock; import java.util.stream.Collectors; public class ExpandSerializer extends JsonSerializer<Object> implements ContextualSerializer { private static final Logger LOG = LoggerFactory.getLogger(ExpandSerializer.class); /** * 成功数据 */ public static final String OK = "OK"; /** * 失败数据 */ public static final String FAIL = "FAIL"; /** * 缓存 */ private static ExpandCache cache; /** * 配置 */ private static JacksonExpandProperties jacksonExpandProperties; /** * 本地锁缓存,防止同时查询 */ private static final LocalCache lockCache = new LocalCache(); /** * 远程调用服务 */ private Object loadService; /** * 方法 */ private String method; /** * 注解参数处理 */ private SerializerParam params; /** * 返回结果处理类 */ private ParamsHandler paramsHandler; /** * 返回结果处理类 */ private ResponseHandler responseHandler; /** * bean名称 */ private String beanName; public ExpandSerializer() { super(); if (cache == null) { synchronized (ExpandSerializer.class) { if (cache == null) { cache = SpringUtil.getBean(ExpandCache.class); jacksonExpandProperties = SpringUtil.getBean(JacksonExpandProperties.class); } } } } public ExpandSerializer(String beanName, String method, SerializerParam params, ParamsHandler paramsHandler, ResponseHandler otherResponseHandler) { this(); this.loadService = SpringUtil.getBean(beanName); this.method = method; this.params = params; this.responseHandler = otherResponseHandler; this.paramsHandler = paramsHandler; this.beanName = beanName; } @Override public void serialize(Object bindData, JsonGenerator gen, SerializerProvider serializers) throws IOException { String writeFieldPath = getFieldPath(gen.getOutputContext()); // 统一path替换 String dynamicExpandCommonPrefix = jacksonExpandProperties.getDynamicExpandCommonPrefix(); if (StringUtils.hasText(dynamicExpandCommonPrefix) && writeFieldPath.startsWith(dynamicExpandCommonPrefix)) { writeFieldPath = writeFieldPath.substring(dynamicExpandCommonPrefix.length() + 1); } // 是否展开 boolean expand; // 动态展开开启,判断是否展开 boolean dynamicExpand = jacksonExpandProperties.isDynamicExpand(); if (dynamicExpand) { Set<String> needExpandField = getParam(jacksonExpandProperties.getDynamicExpandParameterName()); // 如果代码里设置不展开,动态展开也不生效 expand = needExpandField.contains(writeFieldPath) && params.isOpen(); } else { expand = params.isOpen(); } if (!expand) { gen.writeObject(bindData); return; } // 判断要写入的字段 String writeField = gen.getOutputContext().getCurrentName(); if (jacksonExpandProperties.getExpandStrategy() == ExpandStrategy.COVER) { writeField = gen.getOutputContext().getCurrentName(); } else if (jacksonExpandProperties.getExpandStrategy() == ExpandStrategy.COPY) { writeField = String.format(jacksonExpandProperties.getCopyStrategyFormat(), gen.getOutputContext().getCurrentName()); } // 自定义要写入的优先级最高 if (StringUtils.hasText(params.getWriteField())) { writeField = params.getWriteField(); } // 设置理论上的响应类型,要不要使用取决于 ResponseHandler 要不要处理,比如只能写入数据对象存在的对象,默认是忽略存不存在 Class<?> writeClass = null; if (params.getWriteField() != null && StringUtils.hasText(params.getWriteField())) { Field field = ReflectionUtils.findField(gen.getCurrentValue().getClass(), params.getWriteField()); if (field != null) { writeClass = field.getType(); } } // 关闭不存在字段扩展,被写入的字段类型找不到,不扩展 if (!jacksonExpandProperties.isCanExpandToNotExistField() && writeClass == null) { gen.writeObject(bindData); return; } // 翻译为非当前字段,先写入当前字段值再翻译 boolean currField = gen.getOutputContext().getCurrentName().equals(writeField); if (!currField) { gen.writeObject(bindData); gen.writeFieldName(writeField); } if (bindData == null || loadService == null) { gen.writeObject(bindData); return; } // 获取缓存KEY Object[] args = params.getRemoteParams(); int argsLength = args == null ? 0 : args.length; String cacheKey = jacksonExpandProperties.getCachePrefix() + ":" + beanName + ":" + method + ":%s:" + paramsHandler.getCacheKey(bindData, args); Object result = getCacheInfo(cacheKey); if (result != null) { LOG.info("{} Expand cache 命中: {}", beanName, result); gen.writeObject(result); return; } StampedLock
lock = lockCache.get(cacheKey, new StampedLock(), Duration.ofSeconds(300));
// 写锁避免同一业务ID重复查询 long stamp = lock.writeLock(); Integer cacheTime = params.getCacheTime(); try { // 多参数组装 Object[] objectParams = new Object[argsLength + 1]; objectParams[0] = paramsHandler.handleVal(bindData); if(objectParams.length > 1){ System.arraycopy(args, 0, objectParams, 1, argsLength); } // 请求翻译结果 Object loadResult = ReflectUtil.invoke(loadService, method, objectParams); if (loadResult != null) { result = this.responseHandler.handle(this.beanName, method, loadResult, writeClass, objectParams); cache.put(String.format(cacheKey, OK), result, Duration.ofSeconds(cacheTime)); } else { LOG.error("【{}】 Expand失败,未找到:{}", beanName, bindData); cache.put(String.format(cacheKey, FAIL), bindData, Duration.ofSeconds(cacheTime)); result = bindData; } } catch (Exception e) { LOG.error("【{}】 Expand异常:", beanName, e); cache.put(String.format(cacheKey, FAIL), bindData, Duration.ofSeconds(cacheTime)); result = bindData; } finally { lock.unlockWrite(stamp); } gen.writeObject(result); } /** * 获取当前字段的path路径 * * @param outputContext * @return */ private String getFieldPath(JsonStreamContext outputContext) { List<String> path = new ArrayList<>(4); while (outputContext != null) { String currentName = outputContext.getCurrentName(); if (StringUtils.hasText(currentName)) { path.add(currentName); } outputContext = outputContext.getParent(); } Collections.reverse(path); return String.join(".", path); } /** * 获取厍信息 * * @param cacheKey 缓存的KEY * @return */ private Object getCacheInfo(String cacheKey) { Object result = cache.get(String.format(cacheKey, OK)); if (result == null) { result = cache.get(String.format(cacheKey, FAIL)); } return result; } @Override public JsonSerializer<?> createContextual(SerializerProvider prov, BeanProperty property) throws JsonMappingException { if (property != null) { Expand load = property.getAnnotation(Expand.class); if (load == null) { throw new RuntimeException("未注解相关 @Expand 注解"); } String bean = load.bean(); Class<? extends ParamsHandler> paramsHandlerClass = load.paramsHandler(); Class<? extends ResponseHandler> responseHandlerClass = load.responseHandler(); String method = load.method(); try { ParamsHandler paramsHandler = paramsHandlerClass.getDeclaredConstructor().newInstance(); ResponseHandler responseHandler = responseHandlerClass.getDeclaredConstructor().newInstance(); int cacheTime = load.cacheTime(); // 额外参数处理 SerializerParam params = paramsHandler.handleAnnotation(property); // 参数处理器没设置,且父注设置了,以父注解为主 if (params.getCacheTime() == null && cacheTime != -1) { params.setCacheTime(cacheTime); } // 缓存时间未设置,取默认 if (params.getCacheTime() == null) { params.setCacheTime(jacksonExpandProperties.getCacheTimeout()); } if (params.isOpen() == null) { params.setExpand(load.expand()); } return new ExpandSerializer(bean, method, params, paramsHandler, responseHandler); } catch (Exception e) { LOG.error("@Expand error: ", e); } } return prov.findNullValueSerializer(null); } /** * 获取展开参数 * * @param key * @return */ private Set<String> getParam(String key) { RequestAttributes attributes = RequestContextHolder.getRequestAttributes(); if (attributes == null) { return Collections.emptySet(); } ServletRequest request = ((ServletRequestAttributes) attributes).getRequest(); String[] parameterValues = request.getParameterValues(key); if (parameterValues == null) { return Collections.emptySet(); } return Arrays.stream(parameterValues).flatMap(o -> Arrays.stream(o.split(","))) .collect(Collectors.toSet()); } }
src/main/java/com/github/stupdit1t/jackson/expand/serializer/ExpandSerializer.java
stupdit1t-jackson-expand-boot-starter-b492b74
[ { "filename": "src/main/java/com/github/stupdit1t/jackson/expand/cache/ExpandCache.java", "retrieved_chunk": " key.add(String.valueOf(bindData));\n for (Object subVal : annotationVal) {\n key.add(String.valueOf(subVal));\n }\n String cacheKey = properties.getCachePrefix() + \":\" + beanName + \":\" + method + \":%s:\" + key.toString();\n delete(String.format(cacheKey, ExpandSerializer.OK));\n delete(String.format(cacheKey, ExpandSerializer.FAIL));\n }\n /**\n * 模糊匹配key", "score": 0.84365314245224 }, { "filename": "src/main/java/com/github/stupdit1t/jackson/expand/cache/ExpandCache.java", "retrieved_chunk": " *\n * @param key\n */\n void delete(String key);\n /**\n * 按照bean删除缓存\n */\n default void delete(String beanName, String method, Object bindData, Object... annotationVal) {\n JacksonExpandProperties properties = SpringUtil.getBean(JacksonExpandProperties.class);\n StringJoiner key = new StringJoiner(\"-\");", "score": 0.8385152816772461 }, { "filename": "src/main/java/com/github/stupdit1t/jackson/expand/cache/ExpandCache.java", "retrieved_chunk": "package com.github.stupdit1t.jackson.expand.cache;\nimport com.github.stupdit1t.jackson.expand.config.JacksonExpandProperties;\nimport com.github.stupdit1t.jackson.expand.util.SpringUtil;\nimport com.github.stupdit1t.jackson.expand.serializer.ExpandSerializer;\nimport org.springframework.util.AntPathMatcher;\nimport java.time.Duration;\nimport java.util.Set;\nimport java.util.StringJoiner;\n/**\n * 缓存抽象", "score": 0.8270149230957031 }, { "filename": "src/main/java/com/github/stupdit1t/jackson/expand/util/ReflectUtil.java", "retrieved_chunk": " *\n * @param obj\n * @param methodName\n * @param args\n * @return\n */\n public static Object invoke(Object obj, String methodName, Object[] args) throws InvocationTargetException, IllegalAccessException {\n String cacheKey = obj.getClass().getName() + methodName;\n final Method method = METHODS_CACHE.computeIfAbsent(cacheKey, (key) -> getMethod(obj.getClass(), methodName, args));\n if (null == method) {", "score": 0.8259233236312866 }, { "filename": "src/main/java/com/github/stupdit1t/jackson/expand/cache/LocalCache.java", "retrieved_chunk": " * 简单的本地缓存实现\n */\npublic class LocalCache implements ExpandCache {\n private static final Logger LOG = LoggerFactory.getLogger(ExpandSerializer.class);\n private static final Timer timer = new Timer();\n private final Map<String, Object> cacheMap;\n public LocalCache() {\n this.cacheMap = new ConcurrentHashMap<>();\n }\n public LocalCache(Map<String, Object> cacheMap) {", "score": 0.8248964548110962 } ]
java
lock = lockCache.get(cacheKey, new StampedLock(), Duration.ofSeconds(300));
package com.github.stupdit1t.jackson.expand.serializer; import com.fasterxml.jackson.core.JsonGenerator; import com.fasterxml.jackson.core.JsonStreamContext; import com.fasterxml.jackson.databind.BeanProperty; import com.fasterxml.jackson.databind.JsonMappingException; import com.fasterxml.jackson.databind.JsonSerializer; import com.fasterxml.jackson.databind.SerializerProvider; import com.fasterxml.jackson.databind.ser.ContextualSerializer; import com.github.stupdit1t.jackson.expand.annotation.Expand; import com.github.stupdit1t.jackson.expand.cache.ExpandCache; import com.github.stupdit1t.jackson.expand.cache.LocalCache; import com.github.stupdit1t.jackson.expand.config.JacksonExpandProperties; import com.github.stupdit1t.jackson.expand.domain.ExpandStrategy; import com.github.stupdit1t.jackson.expand.domain.SerializerParam; import com.github.stupdit1t.jackson.expand.handler.params.ParamsHandler; import com.github.stupdit1t.jackson.expand.handler.rsp.ResponseHandler; import com.github.stupdit1t.jackson.expand.util.ReflectUtil; import com.github.stupdit1t.jackson.expand.util.SpringUtil; import org.slf4j.Logger; import org.slf4j.LoggerFactory; import org.springframework.util.ReflectionUtils; import org.springframework.util.StringUtils; import org.springframework.web.context.request.RequestAttributes; import org.springframework.web.context.request.RequestContextHolder; import org.springframework.web.context.request.ServletRequestAttributes; import javax.servlet.ServletRequest; import java.io.IOException; import java.lang.reflect.Field; import java.time.Duration; import java.util.*; import java.util.concurrent.locks.StampedLock; import java.util.stream.Collectors; public class ExpandSerializer extends JsonSerializer<Object> implements ContextualSerializer { private static final Logger LOG = LoggerFactory.getLogger(ExpandSerializer.class); /** * 成功数据 */ public static final String OK = "OK"; /** * 失败数据 */ public static final String FAIL = "FAIL"; /** * 缓存 */ private static ExpandCache cache; /** * 配置 */ private static JacksonExpandProperties jacksonExpandProperties; /** * 本地锁缓存,防止同时查询 */ private static final LocalCache lockCache = new LocalCache(); /** * 远程调用服务 */ private Object loadService; /** * 方法 */ private String method; /** * 注解参数处理 */ private SerializerParam params; /** * 返回结果处理类 */ private ParamsHandler paramsHandler; /** * 返回结果处理类 */ private ResponseHandler responseHandler; /** * bean名称 */ private String beanName; public ExpandSerializer() { super(); if (cache == null) { synchronized (ExpandSerializer.class) { if (cache == null) { cache = SpringUtil.getBean(ExpandCache.class); jacksonExpandProperties = SpringUtil.getBean(JacksonExpandProperties.class); } } } } public ExpandSerializer(String beanName, String method, SerializerParam params, ParamsHandler paramsHandler, ResponseHandler otherResponseHandler) { this(); this.loadService = SpringUtil.getBean(beanName); this.method = method; this.params = params; this.responseHandler = otherResponseHandler; this.paramsHandler = paramsHandler; this.beanName = beanName; } @Override public void serialize(Object bindData, JsonGenerator gen, SerializerProvider serializers) throws IOException { String writeFieldPath = getFieldPath(gen.getOutputContext()); // 统一path替换 String dynamicExpandCommonPrefix = jacksonExpandProperties.getDynamicExpandCommonPrefix(); if (StringUtils.hasText(dynamicExpandCommonPrefix) && writeFieldPath.startsWith(dynamicExpandCommonPrefix)) { writeFieldPath = writeFieldPath.substring(dynamicExpandCommonPrefix.length() + 1); } // 是否展开 boolean expand; // 动态展开开启,判断是否展开 boolean dynamicExpand = jacksonExpandProperties.isDynamicExpand(); if (dynamicExpand) { Set<String> needExpandField = getParam(jacksonExpandProperties.getDynamicExpandParameterName()); // 如果代码里设置不展开,动态展开也不生效 expand = needExpandField.contains(writeFieldPath) && params.isOpen(); } else { expand = params.isOpen(); } if (!expand) { gen.writeObject(bindData); return; } // 判断要写入的字段 String writeField = gen.getOutputContext().getCurrentName(); if (jacksonExpandProperties.getExpandStrategy() == ExpandStrategy.COVER) { writeField = gen.getOutputContext().getCurrentName(); } else if (jacksonExpandProperties.getExpandStrategy() == ExpandStrategy.COPY) { writeField = String.
format(jacksonExpandProperties.getCopyStrategyFormat(), gen.getOutputContext().getCurrentName());
} // 自定义要写入的优先级最高 if (StringUtils.hasText(params.getWriteField())) { writeField = params.getWriteField(); } // 设置理论上的响应类型,要不要使用取决于 ResponseHandler 要不要处理,比如只能写入数据对象存在的对象,默认是忽略存不存在 Class<?> writeClass = null; if (params.getWriteField() != null && StringUtils.hasText(params.getWriteField())) { Field field = ReflectionUtils.findField(gen.getCurrentValue().getClass(), params.getWriteField()); if (field != null) { writeClass = field.getType(); } } // 关闭不存在字段扩展,被写入的字段类型找不到,不扩展 if (!jacksonExpandProperties.isCanExpandToNotExistField() && writeClass == null) { gen.writeObject(bindData); return; } // 翻译为非当前字段,先写入当前字段值再翻译 boolean currField = gen.getOutputContext().getCurrentName().equals(writeField); if (!currField) { gen.writeObject(bindData); gen.writeFieldName(writeField); } if (bindData == null || loadService == null) { gen.writeObject(bindData); return; } // 获取缓存KEY Object[] args = params.getRemoteParams(); int argsLength = args == null ? 0 : args.length; String cacheKey = jacksonExpandProperties.getCachePrefix() + ":" + beanName + ":" + method + ":%s:" + paramsHandler.getCacheKey(bindData, args); Object result = getCacheInfo(cacheKey); if (result != null) { LOG.info("{} Expand cache 命中: {}", beanName, result); gen.writeObject(result); return; } StampedLock lock = lockCache.get(cacheKey, new StampedLock(), Duration.ofSeconds(300)); // 写锁避免同一业务ID重复查询 long stamp = lock.writeLock(); Integer cacheTime = params.getCacheTime(); try { // 多参数组装 Object[] objectParams = new Object[argsLength + 1]; objectParams[0] = paramsHandler.handleVal(bindData); if(objectParams.length > 1){ System.arraycopy(args, 0, objectParams, 1, argsLength); } // 请求翻译结果 Object loadResult = ReflectUtil.invoke(loadService, method, objectParams); if (loadResult != null) { result = this.responseHandler.handle(this.beanName, method, loadResult, writeClass, objectParams); cache.put(String.format(cacheKey, OK), result, Duration.ofSeconds(cacheTime)); } else { LOG.error("【{}】 Expand失败,未找到:{}", beanName, bindData); cache.put(String.format(cacheKey, FAIL), bindData, Duration.ofSeconds(cacheTime)); result = bindData; } } catch (Exception e) { LOG.error("【{}】 Expand异常:", beanName, e); cache.put(String.format(cacheKey, FAIL), bindData, Duration.ofSeconds(cacheTime)); result = bindData; } finally { lock.unlockWrite(stamp); } gen.writeObject(result); } /** * 获取当前字段的path路径 * * @param outputContext * @return */ private String getFieldPath(JsonStreamContext outputContext) { List<String> path = new ArrayList<>(4); while (outputContext != null) { String currentName = outputContext.getCurrentName(); if (StringUtils.hasText(currentName)) { path.add(currentName); } outputContext = outputContext.getParent(); } Collections.reverse(path); return String.join(".", path); } /** * 获取厍信息 * * @param cacheKey 缓存的KEY * @return */ private Object getCacheInfo(String cacheKey) { Object result = cache.get(String.format(cacheKey, OK)); if (result == null) { result = cache.get(String.format(cacheKey, FAIL)); } return result; } @Override public JsonSerializer<?> createContextual(SerializerProvider prov, BeanProperty property) throws JsonMappingException { if (property != null) { Expand load = property.getAnnotation(Expand.class); if (load == null) { throw new RuntimeException("未注解相关 @Expand 注解"); } String bean = load.bean(); Class<? extends ParamsHandler> paramsHandlerClass = load.paramsHandler(); Class<? extends ResponseHandler> responseHandlerClass = load.responseHandler(); String method = load.method(); try { ParamsHandler paramsHandler = paramsHandlerClass.getDeclaredConstructor().newInstance(); ResponseHandler responseHandler = responseHandlerClass.getDeclaredConstructor().newInstance(); int cacheTime = load.cacheTime(); // 额外参数处理 SerializerParam params = paramsHandler.handleAnnotation(property); // 参数处理器没设置,且父注设置了,以父注解为主 if (params.getCacheTime() == null && cacheTime != -1) { params.setCacheTime(cacheTime); } // 缓存时间未设置,取默认 if (params.getCacheTime() == null) { params.setCacheTime(jacksonExpandProperties.getCacheTimeout()); } if (params.isOpen() == null) { params.setExpand(load.expand()); } return new ExpandSerializer(bean, method, params, paramsHandler, responseHandler); } catch (Exception e) { LOG.error("@Expand error: ", e); } } return prov.findNullValueSerializer(null); } /** * 获取展开参数 * * @param key * @return */ private Set<String> getParam(String key) { RequestAttributes attributes = RequestContextHolder.getRequestAttributes(); if (attributes == null) { return Collections.emptySet(); } ServletRequest request = ((ServletRequestAttributes) attributes).getRequest(); String[] parameterValues = request.getParameterValues(key); if (parameterValues == null) { return Collections.emptySet(); } return Arrays.stream(parameterValues).flatMap(o -> Arrays.stream(o.split(","))) .collect(Collectors.toSet()); } }
src/main/java/com/github/stupdit1t/jackson/expand/serializer/ExpandSerializer.java
stupdit1t-jackson-expand-boot-starter-b492b74
[ { "filename": "src/main/java/com/github/stupdit1t/jackson/expand/domain/SerializerParam.java", "retrieved_chunk": " * 写入字段\n */\n private String writeField;\n /**\n * 值缓存时间\n */\n private Integer cacheTime;\n /**\n * 是否展开\n */", "score": 0.8399338722229004 }, { "filename": "src/main/java/com/github/stupdit1t/jackson/expand/handler/params/DefaultParamsHandler.java", "retrieved_chunk": " if (expand != null) {\n if (StringUtils.hasText(expand.to())) {\n params.setWriteField(expand.to());\n }\n params.setExpand(expand.expand());\n }\n return params;\n }\n}", "score": 0.8245452642440796 }, { "filename": "src/main/java/com/github/stupdit1t/jackson/expand/domain/ExpandStrategy.java", "retrieved_chunk": "package com.github.stupdit1t.jackson.expand.domain;\n/**\n * 展开策略\n */\npublic enum ExpandStrategy {\n /**\n * 覆盖当前字段\n */\n COVER,\n /**", "score": 0.8192376494407654 }, { "filename": "src/main/java/com/github/stupdit1t/jackson/expand/domain/SerializerParam.java", "retrieved_chunk": " private Boolean expand;\n public Object[] getRemoteParams() {\n return remoteParams;\n }\n public void setRemoteParams(Object[] remoteParams) {\n this.remoteParams = remoteParams;\n }\n public String getWriteField() {\n return writeField;\n }", "score": 0.8112765550613403 }, { "filename": "src/main/java/com/github/stupdit1t/jackson/expand/config/JacksonExpandProperties.java", "retrieved_chunk": " */\n private String dynamicExpandParameterName = \"expand\";\n /**\n * 动态展开 统一数据的Path前缀,比如前缀是 data.body. 如果配置 expand=userId, 相当于是expnad=data.body.userId, 默认无\n */\n private String dynamicExpandCommonPrefix;\n /**\n * 展开策略, 默认覆盖\n */\n private ExpandStrategy expandStrategy = ExpandStrategy.COVER;", "score": 0.8062548637390137 } ]
java
format(jacksonExpandProperties.getCopyStrategyFormat(), gen.getOutputContext().getCurrentName());
package com.github.stupdit1t.jackson.expand.serializer; import com.fasterxml.jackson.core.JsonGenerator; import com.fasterxml.jackson.core.JsonStreamContext; import com.fasterxml.jackson.databind.BeanProperty; import com.fasterxml.jackson.databind.JsonMappingException; import com.fasterxml.jackson.databind.JsonSerializer; import com.fasterxml.jackson.databind.SerializerProvider; import com.fasterxml.jackson.databind.ser.ContextualSerializer; import com.github.stupdit1t.jackson.expand.annotation.Expand; import com.github.stupdit1t.jackson.expand.cache.ExpandCache; import com.github.stupdit1t.jackson.expand.cache.LocalCache; import com.github.stupdit1t.jackson.expand.config.JacksonExpandProperties; import com.github.stupdit1t.jackson.expand.domain.ExpandStrategy; import com.github.stupdit1t.jackson.expand.domain.SerializerParam; import com.github.stupdit1t.jackson.expand.handler.params.ParamsHandler; import com.github.stupdit1t.jackson.expand.handler.rsp.ResponseHandler; import com.github.stupdit1t.jackson.expand.util.ReflectUtil; import com.github.stupdit1t.jackson.expand.util.SpringUtil; import org.slf4j.Logger; import org.slf4j.LoggerFactory; import org.springframework.util.ReflectionUtils; import org.springframework.util.StringUtils; import org.springframework.web.context.request.RequestAttributes; import org.springframework.web.context.request.RequestContextHolder; import org.springframework.web.context.request.ServletRequestAttributes; import javax.servlet.ServletRequest; import java.io.IOException; import java.lang.reflect.Field; import java.time.Duration; import java.util.*; import java.util.concurrent.locks.StampedLock; import java.util.stream.Collectors; public class ExpandSerializer extends JsonSerializer<Object> implements ContextualSerializer { private static final Logger LOG = LoggerFactory.getLogger(ExpandSerializer.class); /** * 成功数据 */ public static final String OK = "OK"; /** * 失败数据 */ public static final String FAIL = "FAIL"; /** * 缓存 */ private static ExpandCache cache; /** * 配置 */ private static JacksonExpandProperties jacksonExpandProperties; /** * 本地锁缓存,防止同时查询 */ private static final LocalCache lockCache = new LocalCache(); /** * 远程调用服务 */ private Object loadService; /** * 方法 */ private String method; /** * 注解参数处理 */ private SerializerParam params; /** * 返回结果处理类 */ private ParamsHandler paramsHandler; /** * 返回结果处理类 */ private ResponseHandler responseHandler; /** * bean名称 */ private String beanName; public ExpandSerializer() { super(); if (cache == null) { synchronized (ExpandSerializer.class) { if (cache == null) { cache = SpringUtil.getBean(ExpandCache.class); jacksonExpandProperties = SpringUtil.getBean(JacksonExpandProperties.class); } } } } public ExpandSerializer(String beanName, String method, SerializerParam params, ParamsHandler paramsHandler, ResponseHandler otherResponseHandler) { this(); this.loadService = SpringUtil.getBean(beanName); this.method = method; this.params = params; this.responseHandler = otherResponseHandler; this.paramsHandler = paramsHandler; this.beanName = beanName; } @Override public void serialize(Object bindData, JsonGenerator gen, SerializerProvider serializers) throws IOException { String writeFieldPath = getFieldPath(gen.getOutputContext()); // 统一path替换 String dynamicExpandCommonPrefix = jacksonExpandProperties.getDynamicExpandCommonPrefix(); if (StringUtils.hasText(dynamicExpandCommonPrefix) && writeFieldPath.startsWith(dynamicExpandCommonPrefix)) { writeFieldPath = writeFieldPath.substring(dynamicExpandCommonPrefix.length() + 1); } // 是否展开 boolean expand; // 动态展开开启,判断是否展开 boolean dynamicExpand = jacksonExpandProperties.isDynamicExpand(); if (dynamicExpand) { Set<String> needExpandField = getParam
(jacksonExpandProperties.getDynamicExpandParameterName());
// 如果代码里设置不展开,动态展开也不生效 expand = needExpandField.contains(writeFieldPath) && params.isOpen(); } else { expand = params.isOpen(); } if (!expand) { gen.writeObject(bindData); return; } // 判断要写入的字段 String writeField = gen.getOutputContext().getCurrentName(); if (jacksonExpandProperties.getExpandStrategy() == ExpandStrategy.COVER) { writeField = gen.getOutputContext().getCurrentName(); } else if (jacksonExpandProperties.getExpandStrategy() == ExpandStrategy.COPY) { writeField = String.format(jacksonExpandProperties.getCopyStrategyFormat(), gen.getOutputContext().getCurrentName()); } // 自定义要写入的优先级最高 if (StringUtils.hasText(params.getWriteField())) { writeField = params.getWriteField(); } // 设置理论上的响应类型,要不要使用取决于 ResponseHandler 要不要处理,比如只能写入数据对象存在的对象,默认是忽略存不存在 Class<?> writeClass = null; if (params.getWriteField() != null && StringUtils.hasText(params.getWriteField())) { Field field = ReflectionUtils.findField(gen.getCurrentValue().getClass(), params.getWriteField()); if (field != null) { writeClass = field.getType(); } } // 关闭不存在字段扩展,被写入的字段类型找不到,不扩展 if (!jacksonExpandProperties.isCanExpandToNotExistField() && writeClass == null) { gen.writeObject(bindData); return; } // 翻译为非当前字段,先写入当前字段值再翻译 boolean currField = gen.getOutputContext().getCurrentName().equals(writeField); if (!currField) { gen.writeObject(bindData); gen.writeFieldName(writeField); } if (bindData == null || loadService == null) { gen.writeObject(bindData); return; } // 获取缓存KEY Object[] args = params.getRemoteParams(); int argsLength = args == null ? 0 : args.length; String cacheKey = jacksonExpandProperties.getCachePrefix() + ":" + beanName + ":" + method + ":%s:" + paramsHandler.getCacheKey(bindData, args); Object result = getCacheInfo(cacheKey); if (result != null) { LOG.info("{} Expand cache 命中: {}", beanName, result); gen.writeObject(result); return; } StampedLock lock = lockCache.get(cacheKey, new StampedLock(), Duration.ofSeconds(300)); // 写锁避免同一业务ID重复查询 long stamp = lock.writeLock(); Integer cacheTime = params.getCacheTime(); try { // 多参数组装 Object[] objectParams = new Object[argsLength + 1]; objectParams[0] = paramsHandler.handleVal(bindData); if(objectParams.length > 1){ System.arraycopy(args, 0, objectParams, 1, argsLength); } // 请求翻译结果 Object loadResult = ReflectUtil.invoke(loadService, method, objectParams); if (loadResult != null) { result = this.responseHandler.handle(this.beanName, method, loadResult, writeClass, objectParams); cache.put(String.format(cacheKey, OK), result, Duration.ofSeconds(cacheTime)); } else { LOG.error("【{}】 Expand失败,未找到:{}", beanName, bindData); cache.put(String.format(cacheKey, FAIL), bindData, Duration.ofSeconds(cacheTime)); result = bindData; } } catch (Exception e) { LOG.error("【{}】 Expand异常:", beanName, e); cache.put(String.format(cacheKey, FAIL), bindData, Duration.ofSeconds(cacheTime)); result = bindData; } finally { lock.unlockWrite(stamp); } gen.writeObject(result); } /** * 获取当前字段的path路径 * * @param outputContext * @return */ private String getFieldPath(JsonStreamContext outputContext) { List<String> path = new ArrayList<>(4); while (outputContext != null) { String currentName = outputContext.getCurrentName(); if (StringUtils.hasText(currentName)) { path.add(currentName); } outputContext = outputContext.getParent(); } Collections.reverse(path); return String.join(".", path); } /** * 获取厍信息 * * @param cacheKey 缓存的KEY * @return */ private Object getCacheInfo(String cacheKey) { Object result = cache.get(String.format(cacheKey, OK)); if (result == null) { result = cache.get(String.format(cacheKey, FAIL)); } return result; } @Override public JsonSerializer<?> createContextual(SerializerProvider prov, BeanProperty property) throws JsonMappingException { if (property != null) { Expand load = property.getAnnotation(Expand.class); if (load == null) { throw new RuntimeException("未注解相关 @Expand 注解"); } String bean = load.bean(); Class<? extends ParamsHandler> paramsHandlerClass = load.paramsHandler(); Class<? extends ResponseHandler> responseHandlerClass = load.responseHandler(); String method = load.method(); try { ParamsHandler paramsHandler = paramsHandlerClass.getDeclaredConstructor().newInstance(); ResponseHandler responseHandler = responseHandlerClass.getDeclaredConstructor().newInstance(); int cacheTime = load.cacheTime(); // 额外参数处理 SerializerParam params = paramsHandler.handleAnnotation(property); // 参数处理器没设置,且父注设置了,以父注解为主 if (params.getCacheTime() == null && cacheTime != -1) { params.setCacheTime(cacheTime); } // 缓存时间未设置,取默认 if (params.getCacheTime() == null) { params.setCacheTime(jacksonExpandProperties.getCacheTimeout()); } if (params.isOpen() == null) { params.setExpand(load.expand()); } return new ExpandSerializer(bean, method, params, paramsHandler, responseHandler); } catch (Exception e) { LOG.error("@Expand error: ", e); } } return prov.findNullValueSerializer(null); } /** * 获取展开参数 * * @param key * @return */ private Set<String> getParam(String key) { RequestAttributes attributes = RequestContextHolder.getRequestAttributes(); if (attributes == null) { return Collections.emptySet(); } ServletRequest request = ((ServletRequestAttributes) attributes).getRequest(); String[] parameterValues = request.getParameterValues(key); if (parameterValues == null) { return Collections.emptySet(); } return Arrays.stream(parameterValues).flatMap(o -> Arrays.stream(o.split(","))) .collect(Collectors.toSet()); } }
src/main/java/com/github/stupdit1t/jackson/expand/serializer/ExpandSerializer.java
stupdit1t-jackson-expand-boot-starter-b492b74
[ { "filename": "src/main/java/com/github/stupdit1t/jackson/expand/handler/params/DefaultParamsHandler.java", "retrieved_chunk": " if (expand != null) {\n if (StringUtils.hasText(expand.to())) {\n params.setWriteField(expand.to());\n }\n params.setExpand(expand.expand());\n }\n return params;\n }\n}", "score": 0.8842716813087463 }, { "filename": "src/main/java/com/github/stupdit1t/jackson/expand/config/JacksonExpandProperties.java", "retrieved_chunk": " */\n private String dynamicExpandParameterName = \"expand\";\n /**\n * 动态展开 统一数据的Path前缀,比如前缀是 data.body. 如果配置 expand=userId, 相当于是expnad=data.body.userId, 默认无\n */\n private String dynamicExpandCommonPrefix;\n /**\n * 展开策略, 默认覆盖\n */\n private ExpandStrategy expandStrategy = ExpandStrategy.COVER;", "score": 0.8676210641860962 }, { "filename": "src/main/java/com/github/stupdit1t/jackson/expand/domain/SerializerParam.java", "retrieved_chunk": " * 写入字段\n */\n private String writeField;\n /**\n * 值缓存时间\n */\n private Integer cacheTime;\n /**\n * 是否展开\n */", "score": 0.8564229011535645 }, { "filename": "src/main/java/com/github/stupdit1t/jackson/expand/domain/SerializerParam.java", "retrieved_chunk": " private Boolean expand;\n public Object[] getRemoteParams() {\n return remoteParams;\n }\n public void setRemoteParams(Object[] remoteParams) {\n this.remoteParams = remoteParams;\n }\n public String getWriteField() {\n return writeField;\n }", "score": 0.8505486249923706 }, { "filename": "src/main/java/com/github/stupdit1t/jackson/expand/config/JacksonExpandProperties.java", "retrieved_chunk": " public boolean isDynamicExpand() {\n return dynamicExpand;\n }\n public void setDynamicExpand(boolean dynamicExpand) {\n this.dynamicExpand = dynamicExpand;\n }\n public String getDynamicExpandParameterName() {\n return dynamicExpandParameterName;\n }\n public void setDynamicExpandParameterName(String dynamicExpandParameterName) {", "score": 0.8465708494186401 } ]
java
(jacksonExpandProperties.getDynamicExpandParameterName());
package com.tati.controller; import org.springframework.beans.factory.annotation.Autowired; import org.springframework.web.bind.annotation.CrossOrigin; import org.springframework.web.bind.annotation.PostMapping; import org.springframework.web.bind.annotation.RequestParam; import org.springframework.web.bind.annotation.RestController; import com.tati.model.YoutubeData; import com.tati.service.YoutubeService; import com.tati.utils.YoutubeAPI; import java.util.List; @CrossOrigin("*") @RestController public class YoutubeAPIController { @Autowired private YoutubeService youtubeService; @PostMapping("/api/data") public YoutubeData add(@RequestParam(name= "url") String url){ if (url != null) { YoutubeAPI youtube = new YoutubeAPI(); List<String> videoData = youtube.getVideoData(url); YoutubeData dataHolder= new YoutubeData(); dataHolder.setVideoTitle(videoData.get(0)); dataHolder.setPublishedAt(videoData.get(1)); dataHolder.setViewCount(videoData.get(2)); dataHolder.setLikeCount(videoData.get(3)); dataHolder.setCommentCount(videoData.get(4)); dataHolder.setEngagementRate(videoData.get(5)); dataHolder.setChannelTitle(videoData.get(6)); dataHolder.setVideoPopularTitle(videoData.get(7));
dataHolder.setVideoPopularLink(videoData.get(8));
youtubeService.save(dataHolder); return dataHolder; } else { return null; } } }
src/main/java/com/tati/controller/YoutubeAPIController.java
tati2002med-dashboarder-spring-mvc-api-61d5a07
[ { "filename": "src/main/java/com/tati/model/YoutubeData.java", "retrieved_chunk": "\tpublic String toString() {\n\t\treturn \"YoutubeData [id=\" + id + \", videoTitle=\" + videoTitle + \", publishedAt=\" + publishedAt + \", viewCount=\"\n\t\t\t\t+ viewCount + \", likeCount=\" + likeCount + \", commentCount=\" + commentCount + \", engagementRate=\"\n\t\t\t\t+ engagementRate + \", channelTitle=\" + channelTitle + \", videoPopularTitle=\" + videoPopularTitle\n\t\t\t\t+ \", videoPopularLink=\" + videoPopularLink + \"]\";\n\t}\n}", "score": 0.8876041173934937 }, { "filename": "src/main/java/com/tati/utils/YoutubeAPI.java", "retrieved_chunk": "\t\t\tString videoPopularTitle = items1.get(0).getSnippet().getTitle();\n\t\t\tString videoPopularId = items1.get(0).getId().getVideoId();\n\t\t\tString videoLink = \"https://www.youtube.com/watch?v=\" + videoPopularId;\n\t\t\t// Title\n\t\t\tString channelTitle = snippet.getChannelTitle();\n\t\t\t// Video Data ----------------------------------\n\t\t\t// Video Title\n\t\t\tString videoTitle = snippet.getTitle();\n\t\t\t// published At\n\t\t\tString publishedTime = snippet.getPublishedAt().toString();", "score": 0.863517165184021 }, { "filename": "src/main/java/com/tati/model/YoutubeData.java", "retrieved_chunk": "\tprivate String videoTitle;\n\tprivate String publishedAt;\n\tprivate String viewCount;\n\tprivate String likeCount;\n\tprivate String commentCount;\n\tprivate String engagementRate;\n\tprivate String channelTitle;\n\tprivate String videoPopularTitle;\n\tprivate String videoPopularLink;\n\tpublic Long getId() {", "score": 0.8521025776863098 }, { "filename": "src/main/java/com/tati/utils/YoutubeAPI.java", "retrieved_chunk": "\t\t\tlist.add(videoTitle.replaceAll(\"&#39;\", \"'\"));\n\t\t\tlist.add(publishedAt);\n\t\t\tlist.add(viewCount);\n\t\t\tlist.add(likeCount);\n\t\t\tlist.add(commentCount);\n\t\t\tlist.add(engagementRate);\n\t\t\tlist.add(channelTitle.replaceAll(\"&#39;\", \"'\"));\n\t\t\tlist.add(videoPopularTitle.replaceAll(\"&#39;\", \"'\"));\n\t\t\tlist.add(videoLink);\n\t\t\treturn list;", "score": 0.851021945476532 }, { "filename": "src/main/java/com/tati/utils/YoutubeAPI.java", "retrieved_chunk": "\t\t\tBigInteger comment_count = statistics.getCommentCount();\n\t\t\tString commentCount = comment_count.toString();\n\t\t\t// Engagement Rate\n\t\t\tBigInteger total_interactions = like_count.add(comment_count);\n\t\t\tDouble engagement_rate = (total_interactions.doubleValue() / view_count.doubleValue()) * 100;\n\t\t\tString engagementRate = String.format(\"%.2f\", engagement_rate);\n\t\t\t// Add data to the list\n\t\t\t// list = [Title, Published At, View Count, Like Count, Comment Count,\n\t\t\t// engagement_rate,Channel name, Most popular video title, Most popular video\n\t\t\t// link]", "score": 0.8471890687942505 } ]
java
dataHolder.setVideoPopularLink(videoData.get(8));
/* * Copyright Strimzi authors. * License: Apache License 2.0 (see the file LICENSE or http://apache.org/licenses/LICENSE-2.0.html). */ package io.strimzi.kafka.bridge.mqtt.core; import io.netty.bootstrap.ServerBootstrap; import io.netty.channel.ChannelFuture; import io.netty.channel.ChannelOption; import io.netty.channel.EventLoopGroup; import io.netty.channel.socket.nio.NioServerSocketChannel; import io.netty.handler.logging.LogLevel; import io.netty.handler.logging.LoggingHandler; import io.strimzi.kafka.bridge.mqtt.config.BridgeConfig; import io.strimzi.kafka.bridge.mqtt.config.MqttConfig; import io.strimzi.kafka.bridge.mqtt.kafka.KafkaBridgeProducer; import org.slf4j.Logger; import org.slf4j.LoggerFactory; /** * Represents the MqttServer component. */ public class MqttServer { private static final Logger logger = LoggerFactory.getLogger(MqttServer.class); private final EventLoopGroup masterGroup; private final EventLoopGroup workerGroup; private final ServerBootstrap serverBootstrap; private final MqttConfig mqttConfig; private final KafkaBridgeProducer kafkaBridgeProducer; private ChannelFuture channelFuture; /** * Constructor * * @param config MqttConfig instance with all configuration needed to run the server. * @param masterGroup EventLoopGroup instance for handle incoming connections. * @param workerGroup EventLoopGroup instance for processing I/O. * @param option ChannelOption<Boolean> instance which allows to configure various channel options, such as SO_KEEPALIVE, SO_BACKLOG etc. * @see BridgeConfig * @see ChannelOption */ public MqttServer(BridgeConfig config, EventLoopGroup masterGroup, EventLoopGroup workerGroup, ChannelOption<Boolean> option) { this.masterGroup = masterGroup; this.workerGroup = workerGroup; this.mqttConfig = config.getMqttConfig(); this.kafkaBridgeProducer = new KafkaBridgeProducer(config.getKafkaConfig()); this.serverBootstrap = new ServerBootstrap(); this.serverBootstrap.group(masterGroup, workerGroup) .channel(NioServerSocketChannel.class) .handler(new LoggingHandler(LogLevel.INFO)) .childHandler(new MqttServerInitializer(this.kafkaBridgeProducer)) .childOption(option, true); } /** * Start the server. */ public void start() throws InterruptedException { // bind the Netty server and wait synchronously this.channelFuture = this.serverBootstrap.bind
(this.mqttConfig.getHost(), this.mqttConfig.getPort()).sync();
} /** * Stop the server. */ public void stop() throws InterruptedException { logger.info("Shutting down Netty server..."); this.channelFuture.channel().close().sync(); this.channelFuture.channel().closeFuture().sync(); this.masterGroup.shutdownGracefully().sync(); this.workerGroup.shutdownGracefully().sync(); logger.info("Netty server shut down"); logger.info("Closing Kafka producers..."); this.kafkaBridgeProducer.close(); logger.info("Kafka producers closed"); } }
src/main/java/io/strimzi/kafka/bridge/mqtt/core/MqttServer.java
strimzi-strimzi-mqtt-bridge-cb0a4b8
[ { "filename": "src/main/java/io/strimzi/kafka/bridge/mqtt/Main.java", "retrieved_chunk": " EventLoopGroup workerGroup = new NioEventLoopGroup();\n MqttServer mqttServer = new MqttServer(bridgeConfig, bossGroup, workerGroup, ChannelOption.SO_KEEPALIVE);\n CountDownLatch latch = new CountDownLatch(1);\n Runtime.getRuntime().addShutdownHook(new Thread(() -> {\n try {\n mqttServer.stop();\n } catch (Exception e) {\n logger.error(\"Error stopping the MQTT server: \", e);\n } finally {\n latch.countDown();", "score": 0.8582572937011719 }, { "filename": "src/test/java/io/strimzi/kafka/bridge/mqtt/MqttBridgetIT.java", "retrieved_chunk": " ));\n // prepare the mapping rules\n String mappingRulesPath = Objects.requireNonNull(MqttBridgetIT.class.getClassLoader().getResource(\"mapping-rules-regex.json\")).getPath();\n MappingRulesLoader.getInstance().init(mappingRulesPath);\n // start the MQTT bridge\n EventLoopGroup bossGroup = new NioEventLoopGroup();\n EventLoopGroup workerGroup = new NioEventLoopGroup();\n mqttBridge = new MqttServer(bridgeConfig, bossGroup, workerGroup, ChannelOption.SO_KEEPALIVE);\n mqttBridge.start();\n }", "score": 0.848699152469635 }, { "filename": "src/main/java/io/strimzi/kafka/bridge/mqtt/Main.java", "retrieved_chunk": " }\n }));\n // start the MQTT server\n mqttServer.start();\n latch.await();\n } catch (Exception e) {\n logger.error(\"Error starting the MQTT server: \", e);\n System.exit(1);\n }\n System.exit(0);", "score": 0.8457454442977905 }, { "filename": "src/main/java/io/strimzi/kafka/bridge/mqtt/core/MqttServerHandler.java", "retrieved_chunk": "public class MqttServerHandler extends SimpleChannelInboundHandler<MqttMessage> {\n private static final Logger logger = LoggerFactory.getLogger(MqttServerHandler.class);\n private final KafkaBridgeProducer kafkaBridgeProducer;\n private MqttKafkaMapper mqttKafkaMapper;\n /**\n * Constructor\n */\n public MqttServerHandler(KafkaBridgeProducer kafkaBridgeProducer) {\n // auto release reference count to avoid memory leak\n super(true);", "score": 0.8159054517745972 }, { "filename": "src/test/java/io/strimzi/kafka/bridge/mqtt/MqttBridgetIT.java", "retrieved_chunk": " private static final String MQTT_SERVER_URI = \"tcp://\" + MQTT_SERVER_HOST + \":\" + MQTT_SERVER_PORT;\n private static final String KAFKA_TOPIC = \"devices_bluetooth_data\";\n private static KafkaConsumer<String, String> kafkaConsumerClient = null;\n private static MqttServer mqttBridge;\n private static StrimziKafkaContainer kafkaContainer;\n /**\n * Start the kafka Cluster\n * Start the MQTT bridge before all tests\n */\n @BeforeAll", "score": 0.8157808184623718 } ]
java
(this.mqttConfig.getHost(), this.mqttConfig.getPort()).sync();
package com.solodroid.ads.sdk.format; import static com.solodroid.ads.sdk.util.Constant.ADMOB; import static com.solodroid.ads.sdk.util.Constant.AD_STATUS_ON; import static com.solodroid.ads.sdk.util.Constant.APPLOVIN; import static com.solodroid.ads.sdk.util.Constant.APPLOVIN_DISCOVERY; import static com.solodroid.ads.sdk.util.Constant.APPLOVIN_MAX; import static com.solodroid.ads.sdk.util.Constant.FACEBOOK; import static com.solodroid.ads.sdk.util.Constant.FAN; import static com.solodroid.ads.sdk.util.Constant.FAN_BIDDING_ADMOB; import static com.solodroid.ads.sdk.util.Constant.FAN_BIDDING_AD_MANAGER; import static com.solodroid.ads.sdk.util.Constant.FAN_BIDDING_APPLOVIN_MAX; import static com.solodroid.ads.sdk.util.Constant.FAN_BIDDING_IRONSOURCE; import static com.solodroid.ads.sdk.util.Constant.GOOGLE_AD_MANAGER; import static com.solodroid.ads.sdk.util.Constant.IRONSOURCE; import static com.solodroid.ads.sdk.util.Constant.MOPUB; import static com.solodroid.ads.sdk.util.Constant.NONE; import static com.solodroid.ads.sdk.util.Constant.STARTAPP; import static com.solodroid.ads.sdk.util.Constant.UNITY; import static com.solodroid.ads.sdk.util.Constant.UNITY_ADS_BANNER_HEIGHT_MEDIUM; import static com.solodroid.ads.sdk.util.Constant.UNITY_ADS_BANNER_WIDTH_MEDIUM; import android.app.Activity; import android.os.Bundle; import android.util.Log; import android.view.View; import android.view.ViewGroup; import android.widget.FrameLayout; import android.widget.RelativeLayout; import androidx.annotation.NonNull; import com.applovin.adview.AppLovinAdView; import com.applovin.mediation.MaxAd; import com.applovin.mediation.MaxAdViewAdListener; import com.applovin.mediation.MaxError; import com.applovin.mediation.ads.MaxAdView; import com.applovin.sdk.AppLovinAd; import com.applovin.sdk.AppLovinAdLoadListener; import com.applovin.sdk.AppLovinAdSize; import com.applovin.sdk.AppLovinSdkUtils; import com.facebook.ads.Ad; import com.facebook.ads.AdSize; import com.google.android.gms.ads.AdListener; import com.google.android.gms.ads.AdRequest; import com.google.android.gms.ads.AdView; import com.google.android.gms.ads.LoadAdError; import com.google.android.gms.ads.admanager.AdManagerAdView; import com.ironsource.mediationsdk.ISBannerSize; import com.ironsource.mediationsdk.IronSource; import com.ironsource.mediationsdk.IronSourceBannerLayout; import com.ironsource.mediationsdk.adunit.adapter.utility.AdInfo; import com.ironsource.mediationsdk.logger.IronSourceError; import com.ironsource.mediationsdk.sdk.LevelPlayBannerListener; import com.solodroid.ads.sdk.R; import com.solodroid.ads.sdk.helper.AppLovinCustomEventBanner; import com.solodroid.ads.sdk.util.Tools; import com.startapp.sdk.ads.banner.Banner; import com.startapp.sdk.ads.banner.BannerListener; import com.unity3d.services.banners.BannerErrorInfo; import com.unity3d.services.banners.BannerView; import com.unity3d.services.banners.UnityBannerSize; public class MediumRectangleAd { public static class Builder { private static final String TAG = "AdNetwork"; private final Activity activity; private AdView adView; private AdManagerAdView adManagerAdView; private com.facebook.ads.AdView fanAdView; private AppLovinAdView appLovinAdView; FrameLayout ironSourceBannerView; private IronSourceBannerLayout ironSourceBannerLayout; private String adStatus = ""; private String adNetwork = ""; private String backupAdNetwork = ""; private String adMobBannerId = ""; private String googleAdManagerBannerId = ""; private String fanBannerId = ""; private String unityBannerId = ""; private String appLovinBannerId = ""; private String appLovinBannerZoneId = ""; private String mopubBannerId = ""; private String ironSourceBannerId = ""; private int placementStatus = 1; private boolean darkTheme = false; private boolean legacyGDPR = false; public Builder(Activity activity) { this.activity = activity; } public Builder build() { loadBannerAd(); return this; } public Builder setAdStatus(String adStatus) { this.adStatus = adStatus; return this; } public Builder setAdNetwork(String adNetwork) { this.adNetwork = adNetwork; return this; } public Builder setBackupAdNetwork(String backupAdNetwork) { this.backupAdNetwork = backupAdNetwork; return this; } public Builder setAdMobBannerId(String adMobBannerId) { this.adMobBannerId = adMobBannerId; return this; } public Builder setGoogleAdManagerBannerId(String googleAdManagerBannerId) { this.googleAdManagerBannerId = googleAdManagerBannerId; return this; } public Builder setFanBannerId(String fanBannerId) { this.fanBannerId = fanBannerId; return this; } public Builder setUnityBannerId(String unityBannerId) { this.unityBannerId = unityBannerId; return this; } public Builder setAppLovinBannerId(String appLovinBannerId) { this.appLovinBannerId = appLovinBannerId; return this; } public Builder setAppLovinBannerZoneId(String appLovinBannerZoneId) { this.appLovinBannerZoneId = appLovinBannerZoneId; return this; } public Builder setMopubBannerId(String mopubBannerId) { this.mopubBannerId = mopubBannerId; return this; } public Builder setIronSourceBannerId(String ironSourceBannerId) { this.ironSourceBannerId = ironSourceBannerId; return this; } public Builder setPlacementStatus(int placementStatus) { this.placementStatus = placementStatus; return this; } public Builder setDarkTheme(boolean darkTheme) { this.darkTheme = darkTheme; return this; } public Builder setLegacyGDPR(boolean legacyGDPR) { this.legacyGDPR = legacyGDPR; return this; } public void loadBannerAd() { if (adStatus.equals(AD_STATUS_ON) && placementStatus != 0) { switch (adNetwork) { case ADMOB: case FAN_BIDDING_ADMOB: FrameLayout adContainerView = activity.findViewById(R.id.admob_banner_view_container); adContainerView.post(() -> { adView = new AdView(activity); adView.setAdUnitId(adMobBannerId); adContainerView.removeAllViews(); adContainerView.addView(adView); adView.setAdSize
(Tools.getAdSizeMREC());
adView.loadAd(Tools.getAdRequest(activity, legacyGDPR)); adView.setAdListener(new AdListener() { @Override public void onAdLoaded() { // Code to be executed when an ad finishes loading. adContainerView.setVisibility(View.VISIBLE); } @Override public void onAdFailedToLoad(@NonNull LoadAdError adError) { // Code to be executed when an ad request fails. adContainerView.setVisibility(View.GONE); loadBackupBannerAd(); } @Override public void onAdOpened() { // Code to be executed when an ad opens an overlay that // covers the screen. } @Override public void onAdClicked() { // Code to be executed when the user clicks on an ad. } @Override public void onAdClosed() { // Code to be executed when the user is about to return // to the app after tapping on an ad. } }); }); Log.d(TAG, adNetwork + " Banner Ad unit Id : " + adMobBannerId); break; case GOOGLE_AD_MANAGER: case FAN_BIDDING_AD_MANAGER: FrameLayout googleAdContainerView = activity.findViewById(R.id.google_ad_banner_view_container); googleAdContainerView.post(() -> { adManagerAdView = new AdManagerAdView(activity); adManagerAdView.setAdUnitId(googleAdManagerBannerId); googleAdContainerView.removeAllViews(); googleAdContainerView.addView(adManagerAdView); adManagerAdView.setAdSize(Tools.getAdSizeMREC()); adManagerAdView.loadAd(Tools.getGoogleAdManagerRequest()); adManagerAdView.setAdListener(new AdListener() { @Override public void onAdClicked() { super.onAdClicked(); } @Override public void onAdClosed() { super.onAdClosed(); } @Override public void onAdFailedToLoad(@NonNull LoadAdError loadAdError) { super.onAdFailedToLoad(loadAdError); googleAdContainerView.setVisibility(View.GONE); loadBackupBannerAd(); } @Override public void onAdImpression() { super.onAdImpression(); } @Override public void onAdLoaded() { super.onAdLoaded(); googleAdContainerView.setVisibility(View.VISIBLE); } @Override public void onAdOpened() { super.onAdOpened(); } }); }); break; case FAN: case FACEBOOK: fanAdView = new com.facebook.ads.AdView(activity, fanBannerId, AdSize.RECTANGLE_HEIGHT_250); RelativeLayout fanAdViewContainer = activity.findViewById(R.id.fan_banner_view_container); fanAdViewContainer.addView(fanAdView); com.facebook.ads.AdListener adListener = new com.facebook.ads.AdListener() { @Override public void onError(Ad ad, com.facebook.ads.AdError adError) { fanAdViewContainer.setVisibility(View.GONE); loadBackupBannerAd(); Log.d(TAG, "Error load FAN : " + adError.getErrorMessage()); } @Override public void onAdLoaded(Ad ad) { fanAdViewContainer.setVisibility(View.VISIBLE); } @Override public void onAdClicked(Ad ad) { } @Override public void onLoggingImpression(Ad ad) { } }; com.facebook.ads.AdView.AdViewLoadConfig loadAdConfig = fanAdView.buildLoadAdConfig().withAdListener(adListener).build(); fanAdView.loadAd(loadAdConfig); break; case STARTAPP: RelativeLayout startAppAdView = activity.findViewById(R.id.startapp_banner_view_container); Banner banner = new Banner(activity, new BannerListener() { @Override public void onReceiveAd(View banner) { startAppAdView.setVisibility(View.VISIBLE); } @Override public void onFailedToReceiveAd(View banner) { startAppAdView.setVisibility(View.GONE); loadBackupBannerAd(); Log.d(TAG, adNetwork + " failed load startapp banner ad : "); } @Override public void onImpression(View view) { } @Override public void onClick(View banner) { } }); startAppAdView.addView(banner); break; case UNITY: RelativeLayout unityAdView = activity.findViewById(R.id.unity_banner_view_container); BannerView bottomBanner = new BannerView(activity, unityBannerId, new UnityBannerSize(UNITY_ADS_BANNER_WIDTH_MEDIUM, UNITY_ADS_BANNER_HEIGHT_MEDIUM)); bottomBanner.setListener(new BannerView.IListener() { @Override public void onBannerLoaded(BannerView bannerView) { unityAdView.setVisibility(View.VISIBLE); Log.d("Unity_banner", "ready"); } @Override public void onBannerShown(BannerView bannerAdView) { } @Override public void onBannerClick(BannerView bannerView) { } @Override public void onBannerFailedToLoad(BannerView bannerView, BannerErrorInfo bannerErrorInfo) { Log.d("SupportTest", "Banner Error" + bannerErrorInfo); unityAdView.setVisibility(View.GONE); loadBackupBannerAd(); } @Override public void onBannerLeftApplication(BannerView bannerView) { } }); unityAdView.addView(bottomBanner); bottomBanner.load(); Log.d(TAG, adNetwork + " Banner Ad unit Id : " + unityBannerId); break; case APPLOVIN: case APPLOVIN_MAX: case FAN_BIDDING_APPLOVIN_MAX: RelativeLayout appLovinAdView = activity.findViewById(R.id.applovin_banner_view_container); MaxAdView maxAdView = new MaxAdView(appLovinBannerId, activity); maxAdView.setListener(new MaxAdViewAdListener() { @Override public void onAdExpanded(MaxAd ad) { } @Override public void onAdCollapsed(MaxAd ad) { } @Override public void onAdLoaded(MaxAd ad) { appLovinAdView.setVisibility(View.VISIBLE); } @Override public void onAdDisplayed(MaxAd ad) { } @Override public void onAdHidden(MaxAd ad) { } @Override public void onAdClicked(MaxAd ad) { } @Override public void onAdLoadFailed(String adUnitId, MaxError error) { appLovinAdView.setVisibility(View.GONE); loadBackupBannerAd(); } @Override public void onAdDisplayFailed(MaxAd ad, MaxError error) { } }); int width = ViewGroup.LayoutParams.MATCH_PARENT; int heightPx = activity.getResources().getDimensionPixelSize(R.dimen.applovin_banner_height); maxAdView.setLayoutParams(new FrameLayout.LayoutParams(width, heightPx)); if (darkTheme) { maxAdView.setBackgroundColor(activity.getResources().getColor(R.color.color_native_background_dark)); } else { maxAdView.setBackgroundColor(activity.getResources().getColor(R.color.color_native_background_light)); } appLovinAdView.addView(maxAdView); maxAdView.loadAd(); Log.d(TAG, adNetwork + " Banner Ad unit Id : " + appLovinBannerId); break; case APPLOVIN_DISCOVERY: RelativeLayout appLovinDiscoveryAdView = activity.findViewById(R.id.applovin_discovery_banner_view_container); AdRequest.Builder builder = new AdRequest.Builder(); Bundle bannerExtras = new Bundle(); bannerExtras.putString("zone_id", appLovinBannerZoneId); builder.addCustomEventExtrasBundle(AppLovinCustomEventBanner.class, bannerExtras); boolean isTablet2 = AppLovinSdkUtils.isTablet(activity); AppLovinAdSize adSize = isTablet2 ? AppLovinAdSize.LEADER : AppLovinAdSize.BANNER; this.appLovinAdView = new AppLovinAdView(adSize, activity); this.appLovinAdView.setAdLoadListener(new AppLovinAdLoadListener() { @Override public void adReceived(AppLovinAd ad) { appLovinDiscoveryAdView.setVisibility(View.VISIBLE); } @Override public void failedToReceiveAd(int errorCode) { appLovinDiscoveryAdView.setVisibility(View.GONE); loadBackupBannerAd(); } }); appLovinDiscoveryAdView.addView(this.appLovinAdView); this.appLovinAdView.loadNextAd(); break; case MOPUB: //Mopub has been acquired by AppLovin break; case IRONSOURCE: case FAN_BIDDING_IRONSOURCE: ironSourceBannerView = activity.findViewById(R.id.ironsource_banner_view_container); ISBannerSize size = ISBannerSize.RECTANGLE; ironSourceBannerLayout = IronSource.createBanner(activity, size); FrameLayout.LayoutParams layoutParams = new FrameLayout.LayoutParams(FrameLayout.LayoutParams.MATCH_PARENT, FrameLayout.LayoutParams.MATCH_PARENT); ironSourceBannerView.addView(ironSourceBannerLayout, 0, layoutParams); if (ironSourceBannerLayout != null) { ironSourceBannerLayout.setLevelPlayBannerListener(new LevelPlayBannerListener() { @Override public void onAdLoaded(AdInfo adInfo) { Log.d(TAG, "onBannerAdLoaded"); ironSourceBannerView.setVisibility(View.VISIBLE); } @Override public void onAdLoadFailed(IronSourceError ironSourceError) { Log.d(TAG, "onBannerAdLoadFailed" + " " + ironSourceError.getErrorMessage()); loadBackupBannerAd(); } @Override public void onAdClicked(AdInfo adInfo) { Log.d(TAG, "onBannerAdClicked"); } @Override public void onAdLeftApplication(AdInfo adInfo) { Log.d(TAG, "onBannerAdLeftApplication"); } @Override public void onAdScreenPresented(AdInfo adInfo) { Log.d(TAG, "onBannerAdScreenPresented"); } @Override public void onAdScreenDismissed(AdInfo adInfo) { Log.d(TAG, "onBannerAdScreenDismissed"); } }); IronSource.loadBanner(ironSourceBannerLayout, ironSourceBannerId); } else { Log.d(TAG, "IronSource.createBanner returned null"); } break; case NONE: //do nothing break; } Log.d(TAG, "Banner Ad is enabled"); } else { Log.d(TAG, "Banner Ad is disabled"); } } public void loadBackupBannerAd() { if (adStatus.equals(AD_STATUS_ON) && placementStatus != 0) { switch (backupAdNetwork) { case ADMOB: case FAN_BIDDING_ADMOB: FrameLayout adContainerView = activity.findViewById(R.id.admob_banner_view_container); adContainerView.post(() -> { adView = new AdView(activity); adView.setAdUnitId(adMobBannerId); adContainerView.removeAllViews(); adContainerView.addView(adView); adView.setAdSize(Tools.getAdSizeMREC()); adView.loadAd(Tools.getAdRequest(activity, legacyGDPR)); adView.setAdListener(new AdListener() { @Override public void onAdLoaded() { // Code to be executed when an ad finishes loading. adContainerView.setVisibility(View.VISIBLE); } @Override public void onAdFailedToLoad(@NonNull LoadAdError adError) { // Code to be executed when an ad request fails. adContainerView.setVisibility(View.GONE); } @Override public void onAdOpened() { // Code to be executed when an ad opens an overlay that // covers the screen. } @Override public void onAdClicked() { // Code to be executed when the user clicks on an ad. } @Override public void onAdClosed() { // Code to be executed when the user is about to return // to the app after tapping on an ad. } }); }); Log.d(TAG, adNetwork + " Banner Ad unit Id : " + adMobBannerId); break; case GOOGLE_AD_MANAGER: case FAN_BIDDING_AD_MANAGER: FrameLayout googleAdContainerView = activity.findViewById(R.id.google_ad_banner_view_container); googleAdContainerView.post(() -> { adManagerAdView = new AdManagerAdView(activity); adManagerAdView.setAdUnitId(googleAdManagerBannerId); googleAdContainerView.removeAllViews(); googleAdContainerView.addView(adManagerAdView); adManagerAdView.setAdSize(Tools.getAdSizeMREC()); adManagerAdView.loadAd(Tools.getGoogleAdManagerRequest()); adManagerAdView.setAdListener(new AdListener() { @Override public void onAdClicked() { super.onAdClicked(); } @Override public void onAdClosed() { super.onAdClosed(); } @Override public void onAdFailedToLoad(@NonNull LoadAdError loadAdError) { super.onAdFailedToLoad(loadAdError); googleAdContainerView.setVisibility(View.GONE); } @Override public void onAdImpression() { super.onAdImpression(); } @Override public void onAdLoaded() { super.onAdLoaded(); googleAdContainerView.setVisibility(View.VISIBLE); } @Override public void onAdOpened() { super.onAdOpened(); } }); }); break; case FAN: case FACEBOOK: fanAdView = new com.facebook.ads.AdView(activity, fanBannerId, AdSize.RECTANGLE_HEIGHT_250); RelativeLayout fanAdViewContainer = activity.findViewById(R.id.fan_banner_view_container); fanAdViewContainer.addView(fanAdView); com.facebook.ads.AdListener adListener = new com.facebook.ads.AdListener() { @Override public void onError(Ad ad, com.facebook.ads.AdError adError) { fanAdViewContainer.setVisibility(View.GONE); Log.d(TAG, "Error load FAN : " + adError.getErrorMessage()); } @Override public void onAdLoaded(Ad ad) { fanAdViewContainer.setVisibility(View.VISIBLE); } @Override public void onAdClicked(Ad ad) { } @Override public void onLoggingImpression(Ad ad) { } }; com.facebook.ads.AdView.AdViewLoadConfig loadAdConfig = fanAdView.buildLoadAdConfig().withAdListener(adListener).build(); fanAdView.loadAd(loadAdConfig); break; case STARTAPP: RelativeLayout startAppAdView = activity.findViewById(R.id.startapp_banner_view_container); Banner banner = new Banner(activity, new BannerListener() { @Override public void onReceiveAd(View banner) { startAppAdView.setVisibility(View.VISIBLE); } @Override public void onFailedToReceiveAd(View banner) { startAppAdView.setVisibility(View.GONE); Log.d(TAG, adNetwork + " failed load startapp banner ad : "); } @Override public void onImpression(View view) { } @Override public void onClick(View banner) { } }); startAppAdView.addView(banner); break; case UNITY: RelativeLayout unityAdView = activity.findViewById(R.id.unity_banner_view_container); BannerView bottomBanner = new BannerView(activity, unityBannerId, new UnityBannerSize(UNITY_ADS_BANNER_WIDTH_MEDIUM, UNITY_ADS_BANNER_HEIGHT_MEDIUM)); bottomBanner.setListener(new BannerView.IListener() { @Override public void onBannerLoaded(BannerView bannerView) { unityAdView.setVisibility(View.VISIBLE); Log.d("Unity_banner", "ready"); } @Override public void onBannerShown(BannerView bannerAdView) { } @Override public void onBannerClick(BannerView bannerView) { } @Override public void onBannerFailedToLoad(BannerView bannerView, BannerErrorInfo bannerErrorInfo) { Log.d("SupportTest", "Banner Error" + bannerErrorInfo); unityAdView.setVisibility(View.GONE); } @Override public void onBannerLeftApplication(BannerView bannerView) { } }); unityAdView.addView(bottomBanner); bottomBanner.load(); Log.d(TAG, adNetwork + " Banner Ad unit Id : " + unityBannerId); break; case APPLOVIN: case APPLOVIN_MAX: case FAN_BIDDING_APPLOVIN_MAX: RelativeLayout appLovinAdView = activity.findViewById(R.id.applovin_banner_view_container); MaxAdView maxAdView = new MaxAdView(appLovinBannerId, activity); maxAdView.setListener(new MaxAdViewAdListener() { @Override public void onAdExpanded(MaxAd ad) { } @Override public void onAdCollapsed(MaxAd ad) { } @Override public void onAdLoaded(MaxAd ad) { appLovinAdView.setVisibility(View.VISIBLE); } @Override public void onAdDisplayed(MaxAd ad) { } @Override public void onAdHidden(MaxAd ad) { } @Override public void onAdClicked(MaxAd ad) { } @Override public void onAdLoadFailed(String adUnitId, MaxError error) { appLovinAdView.setVisibility(View.GONE); } @Override public void onAdDisplayFailed(MaxAd ad, MaxError error) { } }); int width = ViewGroup.LayoutParams.MATCH_PARENT; int heightPx = activity.getResources().getDimensionPixelSize(R.dimen.applovin_banner_height); maxAdView.setLayoutParams(new FrameLayout.LayoutParams(width, heightPx)); if (darkTheme) { maxAdView.setBackgroundColor(activity.getResources().getColor(R.color.color_native_background_dark)); } else { maxAdView.setBackgroundColor(activity.getResources().getColor(R.color.color_native_background_light)); } appLovinAdView.addView(maxAdView); maxAdView.loadAd(); Log.d(TAG, adNetwork + " Banner Ad unit Id : " + appLovinBannerId); break; case APPLOVIN_DISCOVERY: RelativeLayout appLovinDiscoveryAdView = activity.findViewById(R.id.applovin_discovery_banner_view_container); AdRequest.Builder builder = new AdRequest.Builder(); Bundle bannerExtras = new Bundle(); bannerExtras.putString("zone_id", appLovinBannerZoneId); builder.addCustomEventExtrasBundle(AppLovinCustomEventBanner.class, bannerExtras); boolean isTablet2 = AppLovinSdkUtils.isTablet(activity); AppLovinAdSize adSize = isTablet2 ? AppLovinAdSize.LEADER : AppLovinAdSize.BANNER; this.appLovinAdView = new AppLovinAdView(adSize, activity); this.appLovinAdView.setAdLoadListener(new AppLovinAdLoadListener() { @Override public void adReceived(AppLovinAd ad) { appLovinDiscoveryAdView.setVisibility(View.VISIBLE); } @Override public void failedToReceiveAd(int errorCode) { appLovinDiscoveryAdView.setVisibility(View.GONE); } }); appLovinDiscoveryAdView.addView(this.appLovinAdView); this.appLovinAdView.loadNextAd(); break; case MOPUB: //Mopub has been acquired by AppLovin break; case IRONSOURCE: case FAN_BIDDING_IRONSOURCE: ironSourceBannerView = activity.findViewById(R.id.ironsource_banner_view_container); ISBannerSize size = ISBannerSize.RECTANGLE; ironSourceBannerLayout = IronSource.createBanner(activity, size); FrameLayout.LayoutParams layoutParams = new FrameLayout.LayoutParams(FrameLayout.LayoutParams.MATCH_PARENT, FrameLayout.LayoutParams.MATCH_PARENT); ironSourceBannerView.addView(ironSourceBannerLayout, 0, layoutParams); if (ironSourceBannerLayout != null) { ironSourceBannerLayout.setLevelPlayBannerListener(new LevelPlayBannerListener() { @Override public void onAdLoaded(AdInfo adInfo) { Log.d(TAG, "onBannerAdLoaded"); ironSourceBannerView.setVisibility(View.VISIBLE); } @Override public void onAdLoadFailed(IronSourceError ironSourceError) { Log.d(TAG, "onBannerAdLoadFailed" + " " + ironSourceError.getErrorMessage()); } @Override public void onAdClicked(AdInfo adInfo) { Log.d(TAG, "onBannerAdClicked"); } @Override public void onAdLeftApplication(AdInfo adInfo) { Log.d(TAG, "onBannerAdLeftApplication"); } @Override public void onAdScreenPresented(AdInfo adInfo) { Log.d(TAG, "onBannerAdScreenPresented"); } @Override public void onAdScreenDismissed(AdInfo adInfo) { Log.d(TAG, "onBannerAdScreenDismissed"); } }); IronSource.loadBanner(ironSourceBannerLayout, ironSourceBannerId); } else { Log.d(TAG, "IronSource.createBanner returned null"); } break; } Log.d(TAG, "Banner Ad is enabled"); } else { Log.d(TAG, "Banner Ad is disabled"); } } public void destroyAndDetachBanner() { if (adStatus.equals(AD_STATUS_ON) && placementStatus != 0) { if (adNetwork.equals(IRONSOURCE) || backupAdNetwork.equals(IRONSOURCE)) { if (ironSourceBannerView != null) { Log.d(TAG, "ironSource banner is not null, ready to destroy"); IronSource.destroyBanner(ironSourceBannerLayout); ironSourceBannerView.removeView(ironSourceBannerLayout); } else { Log.d(TAG, "ironSource banner is null"); } } } } } }
library/src/main/java/com/solodroid/ads/sdk/format/MediumRectangleAd.java
solodroidx-multi-ads-sdk-4f23cdf
[ { "filename": "library/src/main/java/com/solodroid/ads/sdk/format/InterstitialAd.java", "retrieved_chunk": " case FAN_BIDDING_ADMOB:\n com.google.android.gms.ads.interstitial.InterstitialAd.load(activity, adMobInterstitialId, Tools.getAdRequest(activity, legacyGDPR), new InterstitialAdLoadCallback() {\n @Override\n public void onAdLoaded(@NonNull com.google.android.gms.ads.interstitial.InterstitialAd interstitialAd) {\n adMobInterstitialAd = interstitialAd;\n adMobInterstitialAd.setFullScreenContentCallback(new FullScreenContentCallback() {\n @Override\n public void onAdDismissedFullScreenContent() {\n loadInterstitialAd();\n }", "score": 0.9030089974403381 }, { "filename": "library/src/main/java/com/solodroid/ads/sdk/format/InterstitialAd.java", "retrieved_chunk": " switch (adNetwork) {\n case ADMOB:\n case FAN_BIDDING_ADMOB:\n com.google.android.gms.ads.interstitial.InterstitialAd.load(activity, adMobInterstitialId, Tools.getAdRequest(activity, legacyGDPR), new InterstitialAdLoadCallback() {\n @Override\n public void onAdLoaded(@NonNull com.google.android.gms.ads.interstitial.InterstitialAd interstitialAd) {\n adMobInterstitialAd = interstitialAd;\n adMobInterstitialAd.setFullScreenContentCallback(new FullScreenContentCallback() {\n @Override\n public void onAdDismissedFullScreenContent() {", "score": 0.8945249319076538 }, { "filename": "library/src/main/java/com/solodroid/ads/sdk/format/NativeAd.java", "retrieved_chunk": " startappNativeBackground = activity.findViewById(R.id.startapp_native_background);\n applovinNativeAd = activity.findViewById(R.id.applovin_native_ad_container);\n appLovinDiscoveryMrecAd = activity.findViewById(R.id.applovin_discovery_mrec_ad_container);\n wortiseNativeAd = activity.findViewById(R.id.wortise_native_ad_container);\n switch (adNetwork) {\n case ADMOB:\n case FAN_BIDDING_ADMOB:\n if (admobNativeAd.getVisibility() != View.VISIBLE) {\n AdLoader adLoader = new AdLoader.Builder(activity, adMobNativeId)\n .forNativeAd(NativeAd -> {", "score": 0.8941521048545837 }, { "filename": "library/src/main/java/com/solodroid/ads/sdk/format/NativeAd.java", "retrieved_chunk": " AdOptionsView adOptionsView = new AdOptionsView(activity, fanNativeAd, fanNativeAdLayout);\n adChoicesContainer.removeAllViews();\n adChoicesContainer.addView(adOptionsView, 0);\n // Create native UI using the ad metadata.\n TextView nativeAdTitle = nativeAdView.findViewById(R.id.native_ad_title);\n com.facebook.ads.MediaView nativeAdIcon = nativeAdView.findViewById(R.id.native_ad_icon);\n com.facebook.ads.MediaView nativeAdMedia = nativeAdView.findViewById(R.id.native_ad_media);\n TextView nativeAdSocialContext = nativeAdView.findViewById(R.id.native_ad_social_context);\n TextView nativeAdBody = nativeAdView.findViewById(R.id.native_ad_body);\n TextView sponsoredLabel = nativeAdView.findViewById(R.id.native_ad_sponsored_label);", "score": 0.8936858177185059 }, { "filename": "library/src/main/java/com/solodroid/ads/sdk/format/NativeAdViewHolder.java", "retrieved_chunk": " .build();\n adLoader.loadAd(Tools.getGoogleAdManagerRequest());\n } else {\n Log.d(TAG, \"Ad Manager Native Ad has been loaded\");\n }\n break;\n case FAN:\n case FACEBOOK:\n if (fanNativeAdLayout.getVisibility() != View.VISIBLE) {\n fanNativeAd = new com.facebook.ads.NativeAd(context, fanNativeId);", "score": 0.8929523229598999 } ]
java
(Tools.getAdSizeMREC());
/* * Copyright Strimzi authors. * License: Apache License 2.0 (see the file LICENSE or http://apache.org/licenses/LICENSE-2.0.html). */ package io.strimzi.kafka.bridge.mqtt.config; import java.util.Map; import java.util.stream.Collectors; /** * Represents the bridge configuration properties * * @see MqttConfig * @see KafkaConfig */ public class BridgeConfig extends AbstractConfig { // Prefix for all the specific configuration parameters for the bridge public static final String BRIDGE_CONFIG_PREFIX = "bridge."; // Bridge identification number public static final String BRIDGE_ID = BRIDGE_CONFIG_PREFIX + "id"; private final MqttConfig mqttConfig; private final KafkaConfig kafkaConfig; /** * Constructor * * @param config configuration parameters map * @param mqttConfig MQTT configuration properties * @param kafkaConfig Kafka configuration properties */ public BridgeConfig(Map<String, Object> config, MqttConfig mqttConfig, KafkaConfig kafkaConfig) { super(config); this.mqttConfig = mqttConfig; this.kafkaConfig = kafkaConfig; } /** * Build a bridge configuration object from a map of configuration parameters * * @param map configuration parameters map * @return a new instance of BridgeConfig */ public static BridgeConfig fromMap(Map<String, Object> map) { final MqttConfig
mqttConfig = MqttConfig.fromMap(map);
final KafkaConfig kafkaConfig = KafkaConfig.fromMap(map); return new BridgeConfig(map.entrySet().stream() .filter(entry -> entry.getKey().startsWith(BridgeConfig.BRIDGE_CONFIG_PREFIX)) .collect(Collectors.toMap(Map.Entry::getKey, Map.Entry::getValue)), mqttConfig, kafkaConfig); } /** * @return the Kafka configuration properties */ public KafkaConfig getKafkaConfig() { return this.kafkaConfig; } /** * @return the MQTT configuration properties */ public MqttConfig getMqttConfig() { return this.mqttConfig; } /** * @return the bridge identification number */ public String getBridgeID() { return this.config.get(BridgeConfig.BRIDGE_ID) == null ? null : this.config.get(BridgeConfig.BRIDGE_ID).toString(); } /** * @return the bridge configuration properties */ @Override public String toString() { return "BridgeConfig(" + "config=" + this.config + ", mqttConfig=" + this.mqttConfig + ", kafkaConfig=" + this.kafkaConfig + ")"; } }
src/main/java/io/strimzi/kafka/bridge/mqtt/config/BridgeConfig.java
strimzi-strimzi-mqtt-bridge-cb0a4b8
[ { "filename": "src/main/java/io/strimzi/kafka/bridge/mqtt/config/MqttConfig.java", "retrieved_chunk": " *\n * @param config configuration parameters map\n */\n public MqttConfig(Map<String, Object> config) {\n super(config);\n }\n /**\n * Build a MQTT configuration object from a map of configuration parameters\n *\n * @param map configuration parameters map", "score": 0.8722172975540161 }, { "filename": "src/main/java/io/strimzi/kafka/bridge/mqtt/config/KafkaProducerConfig.java", "retrieved_chunk": " super(config);\n }\n /**\n * Build a Kafka producer configuration object from a map of configuration parameters\n *\n * @param map configuration parameters map\n * @return a new instance of KafkaProducerConfig\n */\n public static KafkaProducerConfig fromMap(Map<String, Object> map) {\n return new KafkaProducerConfig(map.entrySet().stream()", "score": 0.8704622387886047 }, { "filename": "src/test/java/io/strimzi/kafka/bridge/mqtt/config/ConfigTest.java", "retrieved_chunk": " map.put(\"bridge.id\", \"my-bridge\");\n map.put(\"kafka.bootstrap.servers\", \"localhost:9092\");\n map.put(\"kafka.producer.acks\", \"1\");\n map.put(\"mqtt.host\", \"0.0.0.0\");\n map.put(\"mqtt.port\", \"1883\");\n BridgeConfig bridgeConfig = BridgeConfig.fromMap(map);\n assertThat(bridgeConfig.getBridgeID(), is(\"my-bridge\"));\n assertThat(bridgeConfig.getKafkaConfig().getConfig().size(), is(1));\n assertThat(bridgeConfig.getKafkaConfig().getConfig().get(CommonClientConfigs.BOOTSTRAP_SERVERS_CONFIG), is(\"localhost:9092\"));\n assertThat(bridgeConfig.getKafkaConfig().getProducerConfig().getConfig().size(), is(1));", "score": 0.8638039231300354 }, { "filename": "src/main/java/io/strimzi/kafka/bridge/mqtt/config/MqttConfig.java", "retrieved_chunk": " * @return a new instance of MqttConfig\n */\n public static MqttConfig fromMap(Map<String, Object> map) {\n return new MqttConfig(map.entrySet().stream()\n .filter(entry -> entry.getKey().startsWith(MqttConfig.MQTT_CONFIG_PREFIX))\n .collect(Collectors.toMap(Map.Entry::getKey, Map.Entry::getValue)));\n }\n /**\n * @return the MQTT server port\n */", "score": 0.8564655780792236 }, { "filename": "src/main/java/io/strimzi/kafka/bridge/mqtt/config/KafkaConfig.java", "retrieved_chunk": " *\n * @param config configuration parameters map\n * @return a new instance of KafkaConfig\n */\n public static KafkaConfig fromMap(Map<String, Object> config) {\n final KafkaProducerConfig kafkaProducerConfig = KafkaProducerConfig.fromMap(config);\n return new KafkaConfig(config.entrySet().stream()\n .filter((entry -> entry.getKey().startsWith(KafkaConfig.KAFKA_CONFIG_PREFIX) &&\n !entry.getKey().startsWith(KafkaProducerConfig.KAFKA_PRODUCER_CONFIG_PREFIX)))\n .collect(Collectors.toMap((e) -> e.getKey().substring(KAFKA_CONFIG_PREFIX.length()), Map.Entry::getValue)), kafkaProducerConfig);", "score": 0.8494848608970642 } ]
java
mqttConfig = MqttConfig.fromMap(map);
/* * Copyright Strimzi authors. * License: Apache License 2.0 (see the file LICENSE or http://apache.org/licenses/LICENSE-2.0.html). */ package io.strimzi.kafka.bridge.mqtt.core; import io.netty.bootstrap.ServerBootstrap; import io.netty.channel.ChannelFuture; import io.netty.channel.ChannelOption; import io.netty.channel.EventLoopGroup; import io.netty.channel.socket.nio.NioServerSocketChannel; import io.netty.handler.logging.LogLevel; import io.netty.handler.logging.LoggingHandler; import io.strimzi.kafka.bridge.mqtt.config.BridgeConfig; import io.strimzi.kafka.bridge.mqtt.config.MqttConfig; import io.strimzi.kafka.bridge.mqtt.kafka.KafkaBridgeProducer; import org.slf4j.Logger; import org.slf4j.LoggerFactory; /** * Represents the MqttServer component. */ public class MqttServer { private static final Logger logger = LoggerFactory.getLogger(MqttServer.class); private final EventLoopGroup masterGroup; private final EventLoopGroup workerGroup; private final ServerBootstrap serverBootstrap; private final MqttConfig mqttConfig; private final KafkaBridgeProducer kafkaBridgeProducer; private ChannelFuture channelFuture; /** * Constructor * * @param config MqttConfig instance with all configuration needed to run the server. * @param masterGroup EventLoopGroup instance for handle incoming connections. * @param workerGroup EventLoopGroup instance for processing I/O. * @param option ChannelOption<Boolean> instance which allows to configure various channel options, such as SO_KEEPALIVE, SO_BACKLOG etc. * @see BridgeConfig * @see ChannelOption */ public MqttServer(BridgeConfig config, EventLoopGroup masterGroup, EventLoopGroup workerGroup, ChannelOption<Boolean> option) { this.masterGroup = masterGroup; this.workerGroup = workerGroup; this.mqttConfig = config.getMqttConfig(); this.kafkaBridgeProducer = new KafkaBridgeProducer
(config.getKafkaConfig());
this.serverBootstrap = new ServerBootstrap(); this.serverBootstrap.group(masterGroup, workerGroup) .channel(NioServerSocketChannel.class) .handler(new LoggingHandler(LogLevel.INFO)) .childHandler(new MqttServerInitializer(this.kafkaBridgeProducer)) .childOption(option, true); } /** * Start the server. */ public void start() throws InterruptedException { // bind the Netty server and wait synchronously this.channelFuture = this.serverBootstrap.bind(this.mqttConfig.getHost(), this.mqttConfig.getPort()).sync(); } /** * Stop the server. */ public void stop() throws InterruptedException { logger.info("Shutting down Netty server..."); this.channelFuture.channel().close().sync(); this.channelFuture.channel().closeFuture().sync(); this.masterGroup.shutdownGracefully().sync(); this.workerGroup.shutdownGracefully().sync(); logger.info("Netty server shut down"); logger.info("Closing Kafka producers..."); this.kafkaBridgeProducer.close(); logger.info("Kafka producers closed"); } }
src/main/java/io/strimzi/kafka/bridge/mqtt/core/MqttServer.java
strimzi-strimzi-mqtt-bridge-cb0a4b8
[ { "filename": "src/test/java/io/strimzi/kafka/bridge/mqtt/MqttBridgetIT.java", "retrieved_chunk": "/*\n * Copyright Strimzi authors.\n * License: Apache License 2.0 (see the file LICENSE or http://apache.org/licenses/LICENSE-2.0.html).\n */\npackage io.strimzi.kafka.bridge.mqtt;\nimport io.netty.channel.ChannelOption;\nimport io.netty.channel.EventLoopGroup;\nimport io.netty.channel.nio.NioEventLoopGroup;\nimport io.netty.handler.codec.mqtt.MqttQoS;\nimport io.strimzi.kafka.bridge.mqtt.config.BridgeConfig;", "score": 0.8696885108947754 }, { "filename": "src/main/java/io/strimzi/kafka/bridge/mqtt/Main.java", "retrieved_chunk": "/*\n * Copyright Strimzi authors.\n * License: Apache License 2.0 (see the file LICENSE or http://apache.org/licenses/LICENSE-2.0.html).\n */\npackage io.strimzi.kafka.bridge.mqtt;\nimport io.netty.channel.ChannelOption;\nimport io.netty.channel.EventLoopGroup;\nimport io.netty.channel.nio.NioEventLoopGroup;\nimport io.strimzi.kafka.bridge.mqtt.config.BridgeConfig;\nimport io.strimzi.kafka.bridge.mqtt.config.ConfigRetriever;", "score": 0.8546040654182434 }, { "filename": "src/main/java/io/strimzi/kafka/bridge/mqtt/config/BridgeConfig.java", "retrieved_chunk": " /**\n * Constructor\n *\n * @param config configuration parameters map\n * @param mqttConfig MQTT configuration properties\n * @param kafkaConfig Kafka configuration properties\n */\n public BridgeConfig(Map<String, Object> config, MqttConfig mqttConfig, KafkaConfig kafkaConfig) {\n super(config);\n this.mqttConfig = mqttConfig;", "score": 0.8488063812255859 }, { "filename": "src/main/java/io/strimzi/kafka/bridge/mqtt/Main.java", "retrieved_chunk": " EventLoopGroup workerGroup = new NioEventLoopGroup();\n MqttServer mqttServer = new MqttServer(bridgeConfig, bossGroup, workerGroup, ChannelOption.SO_KEEPALIVE);\n CountDownLatch latch = new CountDownLatch(1);\n Runtime.getRuntime().addShutdownHook(new Thread(() -> {\n try {\n mqttServer.stop();\n } catch (Exception e) {\n logger.error(\"Error stopping the MQTT server: \", e);\n } finally {\n latch.countDown();", "score": 0.8192713260650635 }, { "filename": "src/main/java/io/strimzi/kafka/bridge/mqtt/Main.java", "retrieved_chunk": " //load the configuration file from the path specified in the command line\n String configFilePath = getAbsoluteFilePath(cmd.getOptionValue(Main.CONFIG_FILE_OPTION));\n String mappingRulesFile = getAbsoluteFilePath(cmd.getOptionValue(Main.MAPPING_RULES_FILE_OPTION));\n Map<String, ?> configRetriever = configFilePath != null ? ConfigRetriever.getConfig(configFilePath) : ConfigRetriever.getConfigFromEnv();\n BridgeConfig bridgeConfig = BridgeConfig.fromMap((Map<String, Object>) configRetriever);\n logger.info(\"Bridge configuration {}\", bridgeConfig);\n //set the mapping rules file path\n MappingRulesLoader.getInstance().init(mappingRulesFile);\n //start the MQTT server\n EventLoopGroup bossGroup = new NioEventLoopGroup();", "score": 0.8139549493789673 } ]
java
(config.getKafkaConfig());
/* * Copyright Strimzi authors. * License: Apache License 2.0 (see the file LICENSE or http://apache.org/licenses/LICENSE-2.0.html). */ package io.strimzi.kafka.bridge.mqtt.core; import io.netty.bootstrap.ServerBootstrap; import io.netty.channel.ChannelFuture; import io.netty.channel.ChannelOption; import io.netty.channel.EventLoopGroup; import io.netty.channel.socket.nio.NioServerSocketChannel; import io.netty.handler.logging.LogLevel; import io.netty.handler.logging.LoggingHandler; import io.strimzi.kafka.bridge.mqtt.config.BridgeConfig; import io.strimzi.kafka.bridge.mqtt.config.MqttConfig; import io.strimzi.kafka.bridge.mqtt.kafka.KafkaBridgeProducer; import org.slf4j.Logger; import org.slf4j.LoggerFactory; /** * Represents the MqttServer component. */ public class MqttServer { private static final Logger logger = LoggerFactory.getLogger(MqttServer.class); private final EventLoopGroup masterGroup; private final EventLoopGroup workerGroup; private final ServerBootstrap serverBootstrap; private final MqttConfig mqttConfig; private final KafkaBridgeProducer kafkaBridgeProducer; private ChannelFuture channelFuture; /** * Constructor * * @param config MqttConfig instance with all configuration needed to run the server. * @param masterGroup EventLoopGroup instance for handle incoming connections. * @param workerGroup EventLoopGroup instance for processing I/O. * @param option ChannelOption<Boolean> instance which allows to configure various channel options, such as SO_KEEPALIVE, SO_BACKLOG etc. * @see BridgeConfig * @see ChannelOption */ public MqttServer(BridgeConfig config, EventLoopGroup masterGroup, EventLoopGroup workerGroup, ChannelOption<Boolean> option) { this.masterGroup = masterGroup; this.workerGroup = workerGroup; this.mqttConfig =
config.getMqttConfig();
this.kafkaBridgeProducer = new KafkaBridgeProducer(config.getKafkaConfig()); this.serverBootstrap = new ServerBootstrap(); this.serverBootstrap.group(masterGroup, workerGroup) .channel(NioServerSocketChannel.class) .handler(new LoggingHandler(LogLevel.INFO)) .childHandler(new MqttServerInitializer(this.kafkaBridgeProducer)) .childOption(option, true); } /** * Start the server. */ public void start() throws InterruptedException { // bind the Netty server and wait synchronously this.channelFuture = this.serverBootstrap.bind(this.mqttConfig.getHost(), this.mqttConfig.getPort()).sync(); } /** * Stop the server. */ public void stop() throws InterruptedException { logger.info("Shutting down Netty server..."); this.channelFuture.channel().close().sync(); this.channelFuture.channel().closeFuture().sync(); this.masterGroup.shutdownGracefully().sync(); this.workerGroup.shutdownGracefully().sync(); logger.info("Netty server shut down"); logger.info("Closing Kafka producers..."); this.kafkaBridgeProducer.close(); logger.info("Kafka producers closed"); } }
src/main/java/io/strimzi/kafka/bridge/mqtt/core/MqttServer.java
strimzi-strimzi-mqtt-bridge-cb0a4b8
[ { "filename": "src/test/java/io/strimzi/kafka/bridge/mqtt/MqttBridgetIT.java", "retrieved_chunk": "/*\n * Copyright Strimzi authors.\n * License: Apache License 2.0 (see the file LICENSE or http://apache.org/licenses/LICENSE-2.0.html).\n */\npackage io.strimzi.kafka.bridge.mqtt;\nimport io.netty.channel.ChannelOption;\nimport io.netty.channel.EventLoopGroup;\nimport io.netty.channel.nio.NioEventLoopGroup;\nimport io.netty.handler.codec.mqtt.MqttQoS;\nimport io.strimzi.kafka.bridge.mqtt.config.BridgeConfig;", "score": 0.8711925148963928 }, { "filename": "src/main/java/io/strimzi/kafka/bridge/mqtt/Main.java", "retrieved_chunk": "/*\n * Copyright Strimzi authors.\n * License: Apache License 2.0 (see the file LICENSE or http://apache.org/licenses/LICENSE-2.0.html).\n */\npackage io.strimzi.kafka.bridge.mqtt;\nimport io.netty.channel.ChannelOption;\nimport io.netty.channel.EventLoopGroup;\nimport io.netty.channel.nio.NioEventLoopGroup;\nimport io.strimzi.kafka.bridge.mqtt.config.BridgeConfig;\nimport io.strimzi.kafka.bridge.mqtt.config.ConfigRetriever;", "score": 0.8571503758430481 }, { "filename": "src/main/java/io/strimzi/kafka/bridge/mqtt/config/BridgeConfig.java", "retrieved_chunk": " /**\n * Constructor\n *\n * @param config configuration parameters map\n * @param mqttConfig MQTT configuration properties\n * @param kafkaConfig Kafka configuration properties\n */\n public BridgeConfig(Map<String, Object> config, MqttConfig mqttConfig, KafkaConfig kafkaConfig) {\n super(config);\n this.mqttConfig = mqttConfig;", "score": 0.8374571204185486 }, { "filename": "src/main/java/io/strimzi/kafka/bridge/mqtt/Main.java", "retrieved_chunk": " EventLoopGroup workerGroup = new NioEventLoopGroup();\n MqttServer mqttServer = new MqttServer(bridgeConfig, bossGroup, workerGroup, ChannelOption.SO_KEEPALIVE);\n CountDownLatch latch = new CountDownLatch(1);\n Runtime.getRuntime().addShutdownHook(new Thread(() -> {\n try {\n mqttServer.stop();\n } catch (Exception e) {\n logger.error(\"Error stopping the MQTT server: \", e);\n } finally {\n latch.countDown();", "score": 0.832705020904541 }, { "filename": "src/test/java/io/strimzi/kafka/bridge/mqtt/MqttBridgetIT.java", "retrieved_chunk": " ));\n // prepare the mapping rules\n String mappingRulesPath = Objects.requireNonNull(MqttBridgetIT.class.getClassLoader().getResource(\"mapping-rules-regex.json\")).getPath();\n MappingRulesLoader.getInstance().init(mappingRulesPath);\n // start the MQTT bridge\n EventLoopGroup bossGroup = new NioEventLoopGroup();\n EventLoopGroup workerGroup = new NioEventLoopGroup();\n mqttBridge = new MqttServer(bridgeConfig, bossGroup, workerGroup, ChannelOption.SO_KEEPALIVE);\n mqttBridge.start();\n }", "score": 0.8204062581062317 } ]
java
config.getMqttConfig();
/* * Copyright Strimzi authors. * License: Apache License 2.0 (see the file LICENSE or http://apache.org/licenses/LICENSE-2.0.html). */ package io.strimzi.kafka.bridge.mqtt; import io.netty.channel.ChannelOption; import io.netty.channel.EventLoopGroup; import io.netty.channel.nio.NioEventLoopGroup; import io.strimzi.kafka.bridge.mqtt.config.BridgeConfig; import io.strimzi.kafka.bridge.mqtt.config.ConfigRetriever; import io.strimzi.kafka.bridge.mqtt.core.MqttServer; import io.strimzi.kafka.bridge.mqtt.mapper.MappingRulesLoader; import org.apache.commons.cli.CommandLine; import org.apache.commons.cli.DefaultParser; import org.apache.commons.cli.Option; import org.apache.commons.cli.Options; import org.slf4j.Logger; import org.slf4j.LoggerFactory; import java.io.File; import java.util.Map; import java.util.concurrent.CountDownLatch; public class Main { private static final Logger logger = LoggerFactory.getLogger(Main.class); private static final String CONFIG_FILE_OPTION = "config-file"; private static final String MAPPING_RULES_FILE_OPTION = "mapping-rules"; public static void main(String[] args) { logger.info("Strimzi MQTT Bridge {} is starting", Main.class.getPackage().getImplementationVersion()); try { //prepare the command line options CommandLine cmd = new DefaultParser().parse(generateCommandLineOptions(), args); //load the configuration file from the path specified in the command line String configFilePath = getAbsoluteFilePath(cmd.getOptionValue(Main.CONFIG_FILE_OPTION)); String mappingRulesFile = getAbsoluteFilePath(cmd.getOptionValue(Main.MAPPING_RULES_FILE_OPTION)); Map<String, ?> configRetriever = configFilePath != null ? ConfigRetriever.getConfig(configFilePath) : ConfigRetriever.getConfigFromEnv(); BridgeConfig bridgeConfig = BridgeConfig.fromMap((Map<String, Object>) configRetriever); logger.info("Bridge configuration {}", bridgeConfig); //set the mapping rules file path MappingRulesLoader.getInstance().init(mappingRulesFile); //start the MQTT server EventLoopGroup bossGroup = new NioEventLoopGroup(); EventLoopGroup workerGroup = new NioEventLoopGroup(); MqttServer mqttServer = new MqttServer(bridgeConfig, bossGroup, workerGroup, ChannelOption.SO_KEEPALIVE); CountDownLatch latch = new CountDownLatch(1); Runtime.getRuntime().addShutdownHook(new Thread(() -> { try {
mqttServer.stop();
} catch (Exception e) { logger.error("Error stopping the MQTT server: ", e); } finally { latch.countDown(); } })); // start the MQTT server mqttServer.start(); latch.await(); } catch (Exception e) { logger.error("Error starting the MQTT server: ", e); System.exit(1); } System.exit(0); } /** * Generate the command line options. * The options are: * --config-file: the path of the configuration file * --mapping-rules: the path of the topic mapping rules file * E.g.: * <application> --config-file=/path/to/config/file --mapping-rules=/path/to/mapping/rules/file * @return the command line options */ private static Options generateCommandLineOptions() { Options options = new Options(); Option optionConfigFile = Option.builder() .longOpt(Main.CONFIG_FILE_OPTION) .hasArg(true) .required() .desc("The path to the configuration file") .build(); options.addOption(optionConfigFile); Option optionMappingRulesFile = Option.builder() .longOpt(Main.MAPPING_RULES_FILE_OPTION) .hasArg(true) .required() .desc("The path to the topic mapping rules file") .build(); options.addOption(optionMappingRulesFile); return options; } /** * Get the absolute path of the file * * @param arg the path of the file * @return the absolute path of the file */ private static String getAbsoluteFilePath(String arg) { if (arg == null) { return null; } return arg.startsWith(File.separator) ? arg : System.getProperty("user.dir") + File.separator + arg; } }
src/main/java/io/strimzi/kafka/bridge/mqtt/Main.java
strimzi-strimzi-mqtt-bridge-cb0a4b8
[ { "filename": "src/test/java/io/strimzi/kafka/bridge/mqtt/MqttBridgetIT.java", "retrieved_chunk": " ));\n // prepare the mapping rules\n String mappingRulesPath = Objects.requireNonNull(MqttBridgetIT.class.getClassLoader().getResource(\"mapping-rules-regex.json\")).getPath();\n MappingRulesLoader.getInstance().init(mappingRulesPath);\n // start the MQTT bridge\n EventLoopGroup bossGroup = new NioEventLoopGroup();\n EventLoopGroup workerGroup = new NioEventLoopGroup();\n mqttBridge = new MqttServer(bridgeConfig, bossGroup, workerGroup, ChannelOption.SO_KEEPALIVE);\n mqttBridge.start();\n }", "score": 0.9356156587600708 }, { "filename": "src/main/java/io/strimzi/kafka/bridge/mqtt/core/MqttServerHandler.java", "retrieved_chunk": " try {\n MappingRulesLoader mappingRulesLoader = MappingRulesLoader.getInstance();\n List<MappingRule> rules = mappingRulesLoader.loadRules();\n this.mqttKafkaMapper = new MqttKafkaRegexMapper(rules);\n } catch (IOException e) {\n logger.error(\"Error reading mapping file: \", e);\n }\n this.kafkaBridgeProducer = kafkaBridgeProducer;\n }\n /**", "score": 0.8772406578063965 }, { "filename": "src/main/java/io/strimzi/kafka/bridge/mqtt/core/MqttServer.java", "retrieved_chunk": "public class MqttServer {\n private static final Logger logger = LoggerFactory.getLogger(MqttServer.class);\n private final EventLoopGroup masterGroup;\n private final EventLoopGroup workerGroup;\n private final ServerBootstrap serverBootstrap;\n private final MqttConfig mqttConfig;\n private final KafkaBridgeProducer kafkaBridgeProducer;\n private ChannelFuture channelFuture;\n /**\n * Constructor", "score": 0.8533751964569092 }, { "filename": "src/main/java/io/strimzi/kafka/bridge/mqtt/core/MqttServer.java", "retrieved_chunk": " this.workerGroup = workerGroup;\n this.mqttConfig = config.getMqttConfig();\n this.kafkaBridgeProducer = new KafkaBridgeProducer(config.getKafkaConfig());\n this.serverBootstrap = new ServerBootstrap();\n this.serverBootstrap.group(masterGroup, workerGroup)\n .channel(NioServerSocketChannel.class)\n .handler(new LoggingHandler(LogLevel.INFO))\n .childHandler(new MqttServerInitializer(this.kafkaBridgeProducer))\n .childOption(option, true);\n }", "score": 0.8425256609916687 }, { "filename": "src/main/java/io/strimzi/kafka/bridge/mqtt/core/MqttServerHandler.java", "retrieved_chunk": "public class MqttServerHandler extends SimpleChannelInboundHandler<MqttMessage> {\n private static final Logger logger = LoggerFactory.getLogger(MqttServerHandler.class);\n private final KafkaBridgeProducer kafkaBridgeProducer;\n private MqttKafkaMapper mqttKafkaMapper;\n /**\n * Constructor\n */\n public MqttServerHandler(KafkaBridgeProducer kafkaBridgeProducer) {\n // auto release reference count to avoid memory leak\n super(true);", "score": 0.8360405564308167 } ]
java
mqttServer.stop();
/* * Copyright Strimzi authors. * License: Apache License 2.0 (see the file LICENSE or http://apache.org/licenses/LICENSE-2.0.html). */ package io.strimzi.kafka.bridge.mqtt.core; import io.netty.bootstrap.ServerBootstrap; import io.netty.channel.ChannelFuture; import io.netty.channel.ChannelOption; import io.netty.channel.EventLoopGroup; import io.netty.channel.socket.nio.NioServerSocketChannel; import io.netty.handler.logging.LogLevel; import io.netty.handler.logging.LoggingHandler; import io.strimzi.kafka.bridge.mqtt.config.BridgeConfig; import io.strimzi.kafka.bridge.mqtt.config.MqttConfig; import io.strimzi.kafka.bridge.mqtt.kafka.KafkaBridgeProducer; import org.slf4j.Logger; import org.slf4j.LoggerFactory; /** * Represents the MqttServer component. */ public class MqttServer { private static final Logger logger = LoggerFactory.getLogger(MqttServer.class); private final EventLoopGroup masterGroup; private final EventLoopGroup workerGroup; private final ServerBootstrap serverBootstrap; private final MqttConfig mqttConfig; private final KafkaBridgeProducer kafkaBridgeProducer; private ChannelFuture channelFuture; /** * Constructor * * @param config MqttConfig instance with all configuration needed to run the server. * @param masterGroup EventLoopGroup instance for handle incoming connections. * @param workerGroup EventLoopGroup instance for processing I/O. * @param option ChannelOption<Boolean> instance which allows to configure various channel options, such as SO_KEEPALIVE, SO_BACKLOG etc. * @see BridgeConfig * @see ChannelOption */ public MqttServer(BridgeConfig config, EventLoopGroup masterGroup, EventLoopGroup workerGroup, ChannelOption<Boolean> option) { this.masterGroup = masterGroup; this.workerGroup = workerGroup; this.mqttConfig = config.getMqttConfig(); this.kafkaBridgeProducer = new KafkaBridgeProducer(config.getKafkaConfig()); this.serverBootstrap = new ServerBootstrap(); this.serverBootstrap.group(masterGroup, workerGroup) .channel(NioServerSocketChannel.class) .handler(new LoggingHandler(LogLevel.INFO)) .childHandler(new MqttServerInitializer(this.kafkaBridgeProducer)) .childOption(option, true); } /** * Start the server. */ public void start() throws InterruptedException { // bind the Netty server and wait synchronously this.channelFuture = this.serverBootstrap.bind(this.mqttConfig.getHost()
, this.mqttConfig.getPort()).sync();
} /** * Stop the server. */ public void stop() throws InterruptedException { logger.info("Shutting down Netty server..."); this.channelFuture.channel().close().sync(); this.channelFuture.channel().closeFuture().sync(); this.masterGroup.shutdownGracefully().sync(); this.workerGroup.shutdownGracefully().sync(); logger.info("Netty server shut down"); logger.info("Closing Kafka producers..."); this.kafkaBridgeProducer.close(); logger.info("Kafka producers closed"); } }
src/main/java/io/strimzi/kafka/bridge/mqtt/core/MqttServer.java
strimzi-strimzi-mqtt-bridge-cb0a4b8
[ { "filename": "src/main/java/io/strimzi/kafka/bridge/mqtt/Main.java", "retrieved_chunk": " EventLoopGroup workerGroup = new NioEventLoopGroup();\n MqttServer mqttServer = new MqttServer(bridgeConfig, bossGroup, workerGroup, ChannelOption.SO_KEEPALIVE);\n CountDownLatch latch = new CountDownLatch(1);\n Runtime.getRuntime().addShutdownHook(new Thread(() -> {\n try {\n mqttServer.stop();\n } catch (Exception e) {\n logger.error(\"Error stopping the MQTT server: \", e);\n } finally {\n latch.countDown();", "score": 0.8580213189125061 }, { "filename": "src/test/java/io/strimzi/kafka/bridge/mqtt/MqttBridgetIT.java", "retrieved_chunk": " ));\n // prepare the mapping rules\n String mappingRulesPath = Objects.requireNonNull(MqttBridgetIT.class.getClassLoader().getResource(\"mapping-rules-regex.json\")).getPath();\n MappingRulesLoader.getInstance().init(mappingRulesPath);\n // start the MQTT bridge\n EventLoopGroup bossGroup = new NioEventLoopGroup();\n EventLoopGroup workerGroup = new NioEventLoopGroup();\n mqttBridge = new MqttServer(bridgeConfig, bossGroup, workerGroup, ChannelOption.SO_KEEPALIVE);\n mqttBridge.start();\n }", "score": 0.8478949666023254 }, { "filename": "src/main/java/io/strimzi/kafka/bridge/mqtt/Main.java", "retrieved_chunk": " }\n }));\n // start the MQTT server\n mqttServer.start();\n latch.await();\n } catch (Exception e) {\n logger.error(\"Error starting the MQTT server: \", e);\n System.exit(1);\n }\n System.exit(0);", "score": 0.8449010848999023 }, { "filename": "src/test/java/io/strimzi/kafka/bridge/mqtt/MqttBridgetIT.java", "retrieved_chunk": " private static final String MQTT_SERVER_URI = \"tcp://\" + MQTT_SERVER_HOST + \":\" + MQTT_SERVER_PORT;\n private static final String KAFKA_TOPIC = \"devices_bluetooth_data\";\n private static KafkaConsumer<String, String> kafkaConsumerClient = null;\n private static MqttServer mqttBridge;\n private static StrimziKafkaContainer kafkaContainer;\n /**\n * Start the kafka Cluster\n * Start the MQTT bridge before all tests\n */\n @BeforeAll", "score": 0.8153344988822937 }, { "filename": "src/main/java/io/strimzi/kafka/bridge/mqtt/core/MqttServerHandler.java", "retrieved_chunk": "public class MqttServerHandler extends SimpleChannelInboundHandler<MqttMessage> {\n private static final Logger logger = LoggerFactory.getLogger(MqttServerHandler.class);\n private final KafkaBridgeProducer kafkaBridgeProducer;\n private MqttKafkaMapper mqttKafkaMapper;\n /**\n * Constructor\n */\n public MqttServerHandler(KafkaBridgeProducer kafkaBridgeProducer) {\n // auto release reference count to avoid memory leak\n super(true);", "score": 0.8150143623352051 } ]
java
, this.mqttConfig.getPort()).sync();
/* * Copyright Strimzi authors. * License: Apache License 2.0 (see the file LICENSE or http://apache.org/licenses/LICENSE-2.0.html). */ package io.strimzi.kafka.bridge.mqtt; import io.netty.channel.ChannelOption; import io.netty.channel.EventLoopGroup; import io.netty.channel.nio.NioEventLoopGroup; import io.strimzi.kafka.bridge.mqtt.config.BridgeConfig; import io.strimzi.kafka.bridge.mqtt.config.ConfigRetriever; import io.strimzi.kafka.bridge.mqtt.core.MqttServer; import io.strimzi.kafka.bridge.mqtt.mapper.MappingRulesLoader; import org.apache.commons.cli.CommandLine; import org.apache.commons.cli.DefaultParser; import org.apache.commons.cli.Option; import org.apache.commons.cli.Options; import org.slf4j.Logger; import org.slf4j.LoggerFactory; import java.io.File; import java.util.Map; import java.util.concurrent.CountDownLatch; public class Main { private static final Logger logger = LoggerFactory.getLogger(Main.class); private static final String CONFIG_FILE_OPTION = "config-file"; private static final String MAPPING_RULES_FILE_OPTION = "mapping-rules"; public static void main(String[] args) { logger.info("Strimzi MQTT Bridge {} is starting", Main.class.getPackage().getImplementationVersion()); try { //prepare the command line options CommandLine cmd = new DefaultParser().parse(generateCommandLineOptions(), args); //load the configuration file from the path specified in the command line String configFilePath = getAbsoluteFilePath(cmd.getOptionValue(Main.CONFIG_FILE_OPTION)); String mappingRulesFile = getAbsoluteFilePath(cmd.getOptionValue(Main.MAPPING_RULES_FILE_OPTION)); Map<String, ?> configRetriever = configFilePath != null ? ConfigRetriever.getConfig(configFilePath) : ConfigRetriever.getConfigFromEnv(); BridgeConfig bridgeConfig = BridgeConfig.fromMap((Map<String, Object>) configRetriever); logger.info("Bridge configuration {}", bridgeConfig); //set the mapping rules file path
MappingRulesLoader.getInstance().init(mappingRulesFile);
//start the MQTT server EventLoopGroup bossGroup = new NioEventLoopGroup(); EventLoopGroup workerGroup = new NioEventLoopGroup(); MqttServer mqttServer = new MqttServer(bridgeConfig, bossGroup, workerGroup, ChannelOption.SO_KEEPALIVE); CountDownLatch latch = new CountDownLatch(1); Runtime.getRuntime().addShutdownHook(new Thread(() -> { try { mqttServer.stop(); } catch (Exception e) { logger.error("Error stopping the MQTT server: ", e); } finally { latch.countDown(); } })); // start the MQTT server mqttServer.start(); latch.await(); } catch (Exception e) { logger.error("Error starting the MQTT server: ", e); System.exit(1); } System.exit(0); } /** * Generate the command line options. * The options are: * --config-file: the path of the configuration file * --mapping-rules: the path of the topic mapping rules file * E.g.: * <application> --config-file=/path/to/config/file --mapping-rules=/path/to/mapping/rules/file * @return the command line options */ private static Options generateCommandLineOptions() { Options options = new Options(); Option optionConfigFile = Option.builder() .longOpt(Main.CONFIG_FILE_OPTION) .hasArg(true) .required() .desc("The path to the configuration file") .build(); options.addOption(optionConfigFile); Option optionMappingRulesFile = Option.builder() .longOpt(Main.MAPPING_RULES_FILE_OPTION) .hasArg(true) .required() .desc("The path to the topic mapping rules file") .build(); options.addOption(optionMappingRulesFile); return options; } /** * Get the absolute path of the file * * @param arg the path of the file * @return the absolute path of the file */ private static String getAbsoluteFilePath(String arg) { if (arg == null) { return null; } return arg.startsWith(File.separator) ? arg : System.getProperty("user.dir") + File.separator + arg; } }
src/main/java/io/strimzi/kafka/bridge/mqtt/Main.java
strimzi-strimzi-mqtt-bridge-cb0a4b8
[ { "filename": "src/test/java/io/strimzi/kafka/bridge/mqtt/MqttBridgetIT.java", "retrieved_chunk": " ));\n // prepare the mapping rules\n String mappingRulesPath = Objects.requireNonNull(MqttBridgetIT.class.getClassLoader().getResource(\"mapping-rules-regex.json\")).getPath();\n MappingRulesLoader.getInstance().init(mappingRulesPath);\n // start the MQTT bridge\n EventLoopGroup bossGroup = new NioEventLoopGroup();\n EventLoopGroup workerGroup = new NioEventLoopGroup();\n mqttBridge = new MqttServer(bridgeConfig, bossGroup, workerGroup, ChannelOption.SO_KEEPALIVE);\n mqttBridge.start();\n }", "score": 0.8540788888931274 }, { "filename": "src/main/java/io/strimzi/kafka/bridge/mqtt/core/MqttServerHandler.java", "retrieved_chunk": " try {\n MappingRulesLoader mappingRulesLoader = MappingRulesLoader.getInstance();\n List<MappingRule> rules = mappingRulesLoader.loadRules();\n this.mqttKafkaMapper = new MqttKafkaRegexMapper(rules);\n } catch (IOException e) {\n logger.error(\"Error reading mapping file: \", e);\n }\n this.kafkaBridgeProducer = kafkaBridgeProducer;\n }\n /**", "score": 0.8344483971595764 }, { "filename": "src/test/java/io/strimzi/kafka/bridge/mqtt/config/ConfigTest.java", "retrieved_chunk": " map.put(\"bridge.id\", \"my-bridge\");\n map.put(\"kafka.bootstrap.servers\", \"localhost:9092\");\n map.put(\"kafka.producer.acks\", \"1\");\n map.put(\"mqtt.host\", \"0.0.0.0\");\n map.put(\"mqtt.port\", \"1883\");\n BridgeConfig bridgeConfig = BridgeConfig.fromMap(map);\n assertThat(bridgeConfig.getBridgeID(), is(\"my-bridge\"));\n assertThat(bridgeConfig.getKafkaConfig().getConfig().size(), is(1));\n assertThat(bridgeConfig.getKafkaConfig().getConfig().get(CommonClientConfigs.BOOTSTRAP_SERVERS_CONFIG), is(\"localhost:9092\"));\n assertThat(bridgeConfig.getKafkaConfig().getProducerConfig().getConfig().size(), is(1));", "score": 0.8120096921920776 }, { "filename": "src/test/java/io/strimzi/kafka/bridge/mqtt/config/ConfigRetrieverTest.java", "retrieved_chunk": " envs.put(BridgeConfig.BRIDGE_ID, \"my-bridge-env\");\n Map<String, Object> config = ConfigRetriever.getConfig(filePath, envs);\n BridgeConfig bridgeConfig = BridgeConfig.fromMap(config);\n assertThat(\"Bridge-ID should be 'my-bridge-env'\",\n bridgeConfig.getBridgeID(), is(\"my-bridge-env\"));\n }\n}", "score": 0.810735821723938 }, { "filename": "src/test/java/io/strimzi/kafka/bridge/mqtt/config/ConfigRetrieverTest.java", "retrieved_chunk": " @Test\n public void testApplicationPropertiesFile() throws IOException {\n String filePath = Objects.requireNonNull(getClass().getClassLoader().getResource(\"application.properties\")).getPath();\n Map<String, Object> config = ConfigRetriever.getConfig(filePath);\n BridgeConfig bridgeConfig = BridgeConfig.fromMap(config);\n // bridge config related tests\n assertThat(\"Bridge-ID should be 'my-bridge'\",\n bridgeConfig.getBridgeID(), is(\"my-bridge\"));\n // Mqtt server config related tests\n assertThat(\"There should be 2 related mqtt server config parameters\",", "score": 0.8086156845092773 } ]
java
MappingRulesLoader.getInstance().init(mappingRulesFile);
/* * Licensed to the Apache Software Foundation (ASF) under one or more * contributor license agreements. See the NOTICE file distributed with * this work for additional information regarding copyright ownership. * The ASF licenses this file to You under the Apache License, Version 2.0 * (the "License"); you may not use this file except in compliance with * the License. You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ package org.tallison.cc.index.selector; import java.util.HashMap; import java.util.List; import java.util.Map; import com.fasterxml.jackson.annotation.JsonProperty; import org.slf4j.Logger; import org.slf4j.LoggerFactory; import org.tallison.cc.index.CCIndexRecord; public class RecordSelector { private static Logger LOGGER = LoggerFactory.getLogger(RecordSelector.class); public static RecordSelector ACCEPT_ALL_RECORDS = new AcceptAllRecords(); @JsonProperty Map<String, List<SelectorClause>> must = new HashMap<>(); @JsonProperty Map<String, List<SelectorClause>> must_not = new HashMap<>(); @JsonProperty Map<String, List<SelectorClause>> should = new HashMap<>(); public boolean select(CCIndexRecord record) { for (Map.Entry<String, List<SelectorClause>> e : must_not.entrySet()) { String val = getStringValue(e.getKey(), record); if (val == null) { LOGGER.warn("Value is null for '{}' in the must not clause", e.getKey()); continue; } for (SelectorClause clause : e.getValue()) { if (clause.select(val)) { return false; } } } for (Map.Entry<String, List<SelectorClause>> e : must.entrySet()) { String val = getStringValue(e.getKey(), record); if (val == null) { LOGGER.warn("Value is null for '{}' in the must clause. Record not selected.", e.getKey()); return false; } for (SelectorClause clause : e.getValue()) { if (!clause.select(val)) { return false; } } } if (should.size() == 0) { return true; } for (Map.Entry<String, List<SelectorClause>> e : should.entrySet()) { String val = getStringValue(e.getKey(), record); if (val == null) { LOGGER.warn("Value is null for '{}' in the should clause. Record not selected", e.getKey()); continue; } for (SelectorClause clause : e.getValue()) { if (clause.select(val)) { return true; } } } return false; } private String getStringValue(String key, CCIndexRecord record) { switch (key) { case "mime_detected": return record.getMimeDetected(); case "truncated": return record.getTruncated(); case "mime": return record.getMime(); case "status":
return Integer.toString(record.getStatus());
case "url": return record.getUrl(); case "host": return record.getHost(); case "digest": return record.getDigest(); default: throw new IllegalArgumentException("Don't yet support key " + key); } } private static class AcceptAllRecords extends RecordSelector { @Override public boolean select(CCIndexRecord record) { return true; } } }
src/main/java/org/tallison/cc/index/selector/RecordSelector.java
tballison-commoncrawl-fetcher-lite-2c106d1
[ { "filename": "src/main/java/org/tallison/cc/index/CCIndexRecord.java", "retrieved_chunk": " return CCIndexRecord.normalizeMime(mime);\n }\n public String getNormalizedMimeDetected() {\n return CCIndexRecord.normalizeMime(mimeDetected);\n }\n public Integer getStatus() {\n return status;\n }\n public void setStatus(int status) {\n this.status = status;", "score": 0.8390832543373108 }, { "filename": "src/main/java/org/tallison/cc/index/CCIndexRecord.java", "retrieved_chunk": " return \"bytes=\" + offset + \"-\" + (offset + length - 1);\n }\n @Override\n public String toString() {\n return \"CCIndexRecord{\" + \"url='\" + url + '\\'' + \", mime='\" + mime + '\\'' +\n \", mimeDetected='\" + mimeDetected + '\\'' + \", status=\" + status + \", digest='\" +\n digest + '\\'' + \", length=\" + length + \", offset=\" + offset + \", filename='\" +\n filename + '\\'' + \", charset='\" + charset + '\\'' + \", languages='\" + languages +\n '\\'' + \", truncated='\" + truncated + '\\'' + \", redirect='\" + redirect + '\\'' + '}';\n }", "score": 0.8186668753623962 }, { "filename": "src/main/java/org/tallison/cc/index/CCIndexRecord.java", "retrieved_chunk": "import com.fasterxml.jackson.databind.PropertyNamingStrategies;\nimport com.fasterxml.jackson.databind.annotation.JsonNaming;\nimport org.slf4j.Logger;\nimport org.slf4j.LoggerFactory;\n@JsonNaming(PropertyNamingStrategies.KebabCaseStrategy.class)\npublic class CCIndexRecord {\n private static Pattern INT_PATTERN = Pattern.compile(\"^\\\\d+$\");\n private static Logger LOGGER = LoggerFactory.getLogger(CCIndexRecord.class);\n private static ObjectMapper OBJECT_MAPPER = new ObjectMapper();\n private String url;", "score": 0.8135484457015991 }, { "filename": "src/main/java/org/tallison/cc/index/CCIndexRecord.java", "retrieved_chunk": " return \"\";\n }\n public static Optional<CCIndexRecord> parseRecord(String row) {\n int urlI = row.indexOf(' ');\n int dateI = row.indexOf(' ', urlI + 1);\n if (dateI < 0) {\n LOGGER.warn(\"bad record dateI < 0: {}\", row);\n return Optional.empty();\n }\n String json = row.substring(dateI + 1);", "score": 0.7981610298156738 }, { "filename": "src/main/java/org/tallison/cc/index/AbstractRecordProcessor.java", "retrieved_chunk": "import java.util.regex.Pattern;\npublic abstract class AbstractRecordProcessor implements IndexRecordProcessor {\n protected static AtomicInteger threadCounter = new AtomicInteger(0);\n private final int threadNumber;\n public AbstractRecordProcessor() {\n threadNumber = threadCounter.incrementAndGet();\n }\n //returns \"\" if key is null, otherwise, trims and converts remaining \\r\\n\\t to \" \"\n protected static String clean(String key) {\n if (key == null) {", "score": 0.796244740486145 } ]
java
return Integer.toString(record.getStatus());
/* * Licensed to the Apache Software Foundation (ASF) under one or more * contributor license agreements. See the NOTICE file distributed with * this work for additional information regarding copyright ownership. * The ASF licenses this file to You under the Apache License, Version 2.0 * (the "License"); you may not use this file except in compliance with * the License. You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ package org.tallison.cc.index.selector; import java.util.HashMap; import java.util.List; import java.util.Map; import com.fasterxml.jackson.annotation.JsonProperty; import org.slf4j.Logger; import org.slf4j.LoggerFactory; import org.tallison.cc.index.CCIndexRecord; public class RecordSelector { private static Logger LOGGER = LoggerFactory.getLogger(RecordSelector.class); public static RecordSelector ACCEPT_ALL_RECORDS = new AcceptAllRecords(); @JsonProperty Map<String, List<SelectorClause>> must = new HashMap<>(); @JsonProperty Map<String, List<SelectorClause>> must_not = new HashMap<>(); @JsonProperty Map<String, List<SelectorClause>> should = new HashMap<>(); public boolean select(CCIndexRecord record) { for (Map.Entry<String, List<SelectorClause>> e : must_not.entrySet()) { String val = getStringValue(e.getKey(), record); if (val == null) { LOGGER.warn("Value is null for '{}' in the must not clause", e.getKey()); continue; } for (SelectorClause clause : e.getValue()) { if (clause.select(val)) { return false; } } } for (Map.Entry<String, List<SelectorClause>> e : must.entrySet()) { String val = getStringValue(e.getKey(), record); if (val == null) { LOGGER.warn("Value is null for '{}' in the must clause. Record not selected.", e.getKey()); return false; } for (SelectorClause clause : e.getValue()) { if (!clause.select(val)) { return false; } } } if (should.size() == 0) { return true; } for (Map.Entry<String, List<SelectorClause>> e : should.entrySet()) { String val = getStringValue(e.getKey(), record); if (val == null) { LOGGER.warn("Value is null for '{}' in the should clause. Record not selected", e.getKey()); continue; } for (SelectorClause clause : e.getValue()) { if (clause.select(val)) { return true; } } } return false; } private String getStringValue(String key, CCIndexRecord record) { switch (key) { case "mime_detected": return record.getMimeDetected(); case "truncated": return
record.getTruncated();
case "mime": return record.getMime(); case "status": return Integer.toString(record.getStatus()); case "url": return record.getUrl(); case "host": return record.getHost(); case "digest": return record.getDigest(); default: throw new IllegalArgumentException("Don't yet support key " + key); } } private static class AcceptAllRecords extends RecordSelector { @Override public boolean select(CCIndexRecord record) { return true; } } }
src/main/java/org/tallison/cc/index/selector/RecordSelector.java
tballison-commoncrawl-fetcher-lite-2c106d1
[ { "filename": "src/main/java/org/tallison/cc/index/extractor/CCFileExtractorRecordProcessor.java", "retrieved_chunk": " //problem already logged\n return true;\n }\n CCIndexRecord r = record.get();\n if (!fetcherConfig.getRecordSelector().select(r)) {\n return true;\n }\n //if truncated, count appropriately and test for limits\n if (!StringUtils.isBlank(r.getTruncated())) {\n long truncated = counter.getTruncated().incrementAndGet();", "score": 0.8338386416435242 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/CCMimeCounter.java", "retrieved_chunk": " return false;\n }\n //check for hit max\n //return false;\n Optional<CCIndexRecord> record = CCIndexRecord.parseRecord(json);\n if (record.isEmpty()) {\n //problem already logged\n return true;\n }\n CCIndexRecord r = record.get();", "score": 0.8282814621925354 }, { "filename": "src/main/java/org/tallison/cc/index/CCIndexRecord.java", "retrieved_chunk": " return CCIndexRecord.normalizeMime(mime);\n }\n public String getNormalizedMimeDetected() {\n return CCIndexRecord.normalizeMime(mimeDetected);\n }\n public Integer getStatus() {\n return status;\n }\n public void setStatus(int status) {\n this.status = status;", "score": 0.8279832601547241 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/CCMimeCounter.java", "retrieved_chunk": " return true;\n }\n }\n private static class DetectedMimeCounter extends AbstractRecordProcessor {\n private final ExtractorConfig fetcherConfig;\n private final CCIndexReaderCounter counter;\n private final Map<String, MutableLong> totalCounts = new HashMap<>();\n private final Map<String, MutableLong> truncatedCounts = new HashMap<>();\n public DetectedMimeCounter(ExtractorConfig fetcherConfig, CCIndexReaderCounter counter) {\n this.fetcherConfig = fetcherConfig;", "score": 0.8116198778152466 }, { "filename": "src/main/java/org/tallison/cc/index/selector/AcceptAllRecords.java", "retrieved_chunk": " public boolean select(CCIndexRecord record) {\n return true;\n }\n}", "score": 0.8112835884094238 } ]
java
record.getTruncated();
/* * Licensed to the Apache Software Foundation (ASF) under one or more * contributor license agreements. See the NOTICE file distributed with * this work for additional information regarding copyright ownership. * The ASF licenses this file to You under the Apache License, Version 2.0 * (the "License"); you may not use this file except in compliance with * the License. You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ package org.tallison.cc.index.extractor; import java.nio.file.Files; import java.nio.file.Paths; /** * This is a lighter class that doesn't rely on a database * to extract files from CC and write a list of truncated urls. */ public class CCFetcherCli { public static void main(String[] args) throws Exception { String command = args[0]; if (command.equals("Fetch")) { CCFileExtractor.main(new String[]{args[1]}); } else if (command.equals("FetchIndices")) { CCIndexFetcher.main(new String[]{args[1]}); } else if (command.equals("CountMimes")) {
CCMimeCounter.main(new String[]{
args[1]}); } else if (Files.isRegularFile(Paths.get(command))) { CCFileExtractor.main(new String[]{args[0]}); } else { System.out.println("Must start with a command: Fetch, FetchIndices or CountMimes"); } } }
src/main/java/org/tallison/cc/index/extractor/CCFetcherCli.java
tballison-commoncrawl-fetcher-lite-2c106d1
[ { "filename": "src/main/java/org/tallison/cc/index/extractor/CCMimeCounter.java", "retrieved_chunk": " private static final Logger LOGGER = LoggerFactory.getLogger(CCMimeCounter.class);\n public static void main(String[] args) throws Exception {\n ExtractorConfig fetcherConfig =\n new ObjectMapper().readValue(new File(args[0]), ExtractorConfig.class);\n execute(fetcherConfig);\n }\n private static void execute(ExtractorConfig fetcherConfig) throws IOException, TikaException {\n ArrayBlockingQueue<FetchEmitTuple> indexPathsList = new ArrayBlockingQueue<>(1000);\n //IndexPathsReader reads a file containing a list of cc-index.paths files\n //and writes the literal gz files (cc-index/collections/CC-MAIN-2023-06/indexes/cdx-00000.gz)", "score": 0.9018943905830383 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/CCFileExtractor.java", "retrieved_chunk": "import org.apache.tika.utils.StringUtils;\n/**\n * This is a lighter class that doesn't rely on a database\n * to extract files from CC and log a list of truncated urls.\n */\npublic class CCFileExtractor {\n private static final Long INDEX_WORKER_ID = 1l;\n private static final Long INDEX_READER_ID = 2l;\n private static final Logger LOGGER = LoggerFactory.getLogger(CCFileExtractor.class);\n public static void main(String[] args) throws Exception {", "score": 0.8725379109382629 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/CCIndexFetcher.java", "retrieved_chunk": "import org.apache.tika.pipes.pipesiterator.PipesIterator;\n/**\n * This class fetches index files from aws to a local file share.\n * <p>\n * This pulls the index files either via https or s3\n */\npublic class CCIndexFetcher {\n private static final Logger LOGGER = LoggerFactory.getLogger(CCIndexFetcher.class);\n public static void main(String[] args) throws Exception {\n ExtractorConfig fetcherConfig =", "score": 0.8673886060714722 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/CCIndexFetcher.java", "retrieved_chunk": " new ObjectMapper().readValue(new File(args[0]), ExtractorConfig.class);\n execute(fetcherConfig);\n }\n private static void execute(ExtractorConfig fetcherConfig) throws Exception {\n ArrayBlockingQueue<FetchEmitTuple> indexPathsList = new ArrayBlockingQueue<>(1000);\n //IndexPathsReader reads a file containing a list of cc-index.paths files\n //and writes the literal gz files (cc-index/collections/CC-MAIN-2023-06/indexes/cdx-00000.gz)\n //to indexPathsList\n int totalThreads = fetcherConfig.getNumThreads() + 1;\n ExecutorService executorService = Executors.newFixedThreadPool(totalThreads);", "score": 0.8652521967887878 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/CCFileExtractor.java", "retrieved_chunk": " ExtractorConfig fetcherConfig =\n new ObjectMapper().readValue(new File(args[0]), ExtractorConfig.class);\n execute(fetcherConfig);\n }\n private static void execute(ExtractorConfig fetcherConfig) throws TikaException {\n ArrayBlockingQueue<FetchEmitTuple> indexPathsList = new ArrayBlockingQueue<>(1000);\n //IndexPathsReader reads a file containing a list of cc-index.paths files\n //and writes the literal gz files (cc-index/collections/CC-MAIN-2023-06/indexes/cdx-00000.gz)\n //to indexPathsList\n //IndexWorker reads a single index.gz file at a time and processes each record", "score": 0.8619921207427979 } ]
java
CCMimeCounter.main(new String[]{
/* * Licensed to the Apache Software Foundation (ASF) under one or more * contributor license agreements. See the NOTICE file distributed with * this work for additional information regarding copyright ownership. * The ASF licenses this file to You under the Apache License, Version 2.0 * (the "License"); you may not use this file except in compliance with * the License. You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ package org.tallison.cc.index.extractor; import java.nio.file.Files; import java.nio.file.Paths; /** * This is a lighter class that doesn't rely on a database * to extract files from CC and write a list of truncated urls. */ public class CCFetcherCli { public static void main(String[] args) throws Exception { String command = args[0]; if (command.equals("Fetch")) {
CCFileExtractor.main(new String[]{
args[1]}); } else if (command.equals("FetchIndices")) { CCIndexFetcher.main(new String[]{args[1]}); } else if (command.equals("CountMimes")) { CCMimeCounter.main(new String[]{args[1]}); } else if (Files.isRegularFile(Paths.get(command))) { CCFileExtractor.main(new String[]{args[0]}); } else { System.out.println("Must start with a command: Fetch, FetchIndices or CountMimes"); } } }
src/main/java/org/tallison/cc/index/extractor/CCFetcherCli.java
tballison-commoncrawl-fetcher-lite-2c106d1
[ { "filename": "src/main/java/org/tallison/cc/index/extractor/CCFileExtractor.java", "retrieved_chunk": "import org.apache.tika.utils.StringUtils;\n/**\n * This is a lighter class that doesn't rely on a database\n * to extract files from CC and log a list of truncated urls.\n */\npublic class CCFileExtractor {\n private static final Long INDEX_WORKER_ID = 1l;\n private static final Long INDEX_READER_ID = 2l;\n private static final Logger LOGGER = LoggerFactory.getLogger(CCFileExtractor.class);\n public static void main(String[] args) throws Exception {", "score": 0.8927613496780396 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/CCIndexFetcher.java", "retrieved_chunk": "import org.apache.tika.pipes.pipesiterator.PipesIterator;\n/**\n * This class fetches index files from aws to a local file share.\n * <p>\n * This pulls the index files either via https or s3\n */\npublic class CCIndexFetcher {\n private static final Logger LOGGER = LoggerFactory.getLogger(CCIndexFetcher.class);\n public static void main(String[] args) throws Exception {\n ExtractorConfig fetcherConfig =", "score": 0.8674898147583008 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/CCMimeCounter.java", "retrieved_chunk": " private static final Logger LOGGER = LoggerFactory.getLogger(CCMimeCounter.class);\n public static void main(String[] args) throws Exception {\n ExtractorConfig fetcherConfig =\n new ObjectMapper().readValue(new File(args[0]), ExtractorConfig.class);\n execute(fetcherConfig);\n }\n private static void execute(ExtractorConfig fetcherConfig) throws IOException, TikaException {\n ArrayBlockingQueue<FetchEmitTuple> indexPathsList = new ArrayBlockingQueue<>(1000);\n //IndexPathsReader reads a file containing a list of cc-index.paths files\n //and writes the literal gz files (cc-index/collections/CC-MAIN-2023-06/indexes/cdx-00000.gz)", "score": 0.8596946001052856 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/CCIndexFetcher.java", "retrieved_chunk": " new ObjectMapper().readValue(new File(args[0]), ExtractorConfig.class);\n execute(fetcherConfig);\n }\n private static void execute(ExtractorConfig fetcherConfig) throws Exception {\n ArrayBlockingQueue<FetchEmitTuple> indexPathsList = new ArrayBlockingQueue<>(1000);\n //IndexPathsReader reads a file containing a list of cc-index.paths files\n //and writes the literal gz files (cc-index/collections/CC-MAIN-2023-06/indexes/cdx-00000.gz)\n //to indexPathsList\n int totalThreads = fetcherConfig.getNumThreads() + 1;\n ExecutorService executorService = Executors.newFixedThreadPool(totalThreads);", "score": 0.8405793309211731 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/CCFileExtractor.java", "retrieved_chunk": " ExtractorConfig fetcherConfig =\n new ObjectMapper().readValue(new File(args[0]), ExtractorConfig.class);\n execute(fetcherConfig);\n }\n private static void execute(ExtractorConfig fetcherConfig) throws TikaException {\n ArrayBlockingQueue<FetchEmitTuple> indexPathsList = new ArrayBlockingQueue<>(1000);\n //IndexPathsReader reads a file containing a list of cc-index.paths files\n //and writes the literal gz files (cc-index/collections/CC-MAIN-2023-06/indexes/cdx-00000.gz)\n //to indexPathsList\n //IndexWorker reads a single index.gz file at a time and processes each record", "score": 0.8357306718826294 } ]
java
CCFileExtractor.main(new String[]{
package com.suimz.open.chatgptweb.java.core.exception; import com.suimz.open.chatgptweb.java.bean.resp.R; import org.springframework.http.ResponseEntity; import org.springframework.http.converter.HttpMessageNotReadableException; import org.springframework.validation.BindException; import org.springframework.validation.FieldError; import org.springframework.web.bind.MethodArgumentNotValidException; import org.springframework.web.bind.annotation.ExceptionHandler; import org.springframework.web.bind.annotation.RestControllerAdvice; import org.springframework.web.method.annotation.MethodArgumentTypeMismatchException; import java.util.ArrayList; import java.util.HashMap; import java.util.List; import java.util.Map; /** * Handle Exceptions * * @author https://github.com/suimz */ @RestControllerAdvice public class AdviceException { @ExceptionHandler(BizException.class) public ResponseEntity<R> map(BizException e) { return ResponseEntity.status(e.getHttpStatus()).body(R.error(e.getMessage())); } @ExceptionHandler(UnauthorizedBizException.class) public R map(UnauthorizedBizException e) { return R.builder() .status("Unauthorized") .message(e.getMessage()) .build(); } @ExceptionHandler(HttpMessageNotReadableException.class) public R map(HttpMessageNotReadableException e) { return
R.error("bad request");
} @ExceptionHandler(BindException.class) public R map(BindException e) { List<FieldError> fieldErrors = e.getBindingResult().getFieldErrors(); Map<String, Object> error = this.getValidError(fieldErrors); return R.error(error.get("errorMsg").toString()); } @ExceptionHandler(MethodArgumentNotValidException.class) public R map(MethodArgumentNotValidException e) { List<FieldError> fieldErrors = e.getBindingResult().getFieldErrors(); Map<String, Object> error = this.getValidError(fieldErrors); return R.error(error.get("errorMsg").toString()); } private Map<String, Object> getValidError(List<FieldError> fieldErrors) { Map<String, Object> map = new HashMap<String, Object>(16); List<String> errorList = new ArrayList<String>(); StringBuffer errorMsg = new StringBuffer(); for (FieldError error : fieldErrors) { errorList.add(error.getDefaultMessage()); errorMsg.append(error.getDefaultMessage()); // first break; } map.put("errorList", errorList); map.put("errorMsg", errorMsg); return map; } @ExceptionHandler(MethodArgumentTypeMismatchException.class) public R map(MethodArgumentTypeMismatchException e) { return R.error(e.getMessage()); } @ExceptionHandler(Exception.class) public R map(Exception e) { return R.error(); } }
src/main/java/com/suimz/open/chatgptweb/java/core/exception/AdviceException.java
suimz-chatgpt-web-java-37a4c63
[ { "filename": "src/main/java/com/suimz/open/chatgptweb/java/controller/AppController.java", "retrieved_chunk": " thrEx = (BizException) e;\n } else if (e instanceof OpenAiHttpException) {\n OpenAiHttpException exception = (OpenAiHttpException) e;\n thrEx = new ApiRequestErrorBizException(exception.statusCode, exception.getMessage());\n } else if (e.getCause() instanceof SocketTimeoutException) {\n thrEx = new ApiRequestErrorBizException(0);\n } else if (e.getCause() instanceof SocketException || e.getCause() instanceof IOException) {\n thrEx = new ApiRequestErrorBizException(-1);\n } else {\n thrEx = new ApiRequestErrorBizException();", "score": 0.8383499383926392 }, { "filename": "src/main/java/com/suimz/open/chatgptweb/java/core/exception/UnauthorizedBizException.java", "retrieved_chunk": "package com.suimz.open.chatgptweb.java.core.exception;\nimport org.springframework.http.HttpStatus;\n/**\n * Unauthorized Exception\n *\n * @author https://github.com/suimz\n */\npublic class UnauthorizedBizException extends BizException {\n public UnauthorizedBizException() {\n super(HttpStatus.UNAUTHORIZED, \"Error: 无访问权限 | No access rights\");", "score": 0.8273347020149231 }, { "filename": "src/main/java/com/suimz/open/chatgptweb/java/bean/po/ReverseResponseBodyCallback.java", "retrieved_chunk": " ResponseBody errorBody = response.errorBody();\n if (errorBody == null) {\n throw e;\n } else {\n OpenAiError error = mapper.readValue(\n errorBody.string(),\n OpenAiError.class\n );\n throw new OpenAiHttpException(error, e, e.code());\n }", "score": 0.8149840235710144 }, { "filename": "src/main/java/com/suimz/open/chatgptweb/java/bean/resp/R.java", "retrieved_chunk": " public static <T> R<T> ok(T data) {\n return R.ok(null, data);\n }\n public static <T> R<T> ok(String message, T data) {\n return (R<T>) R.builder().status(\"Success\").message(message).data(data).build();\n }\n public static R error() {\n return R.error(\"服务异常 | server exception\");\n }\n public static R error(String error) {", "score": 0.8121318221092224 }, { "filename": "src/main/java/com/suimz/open/chatgptweb/java/core/exception/ApiRequestErrorBizException.java", "retrieved_chunk": "package com.suimz.open.chatgptweb.java.core.exception;\nimport java.util.HashMap;\nimport java.util.Map;\npublic class ApiRequestErrorBizException extends BizException {\n public static final Map<Integer, String> OPENAI_HTTP_ERROR;\n static {\n OPENAI_HTTP_ERROR = new HashMap<>();\n OPENAI_HTTP_ERROR.put(-1, \"[OpenAI] 发送请求失败 | Failed to send request\");\n OPENAI_HTTP_ERROR.put(0, \"[OpenAI] 请求超时 | Request timeout\");\n OPENAI_HTTP_ERROR.put(401, \"[OpenAI] 提供错误的API密钥 | Incorrect API key provided\");", "score": 0.8044658899307251 } ]
java
R.error("bad request");
package com.suimz.open.chatgptweb.java.core.exception; import com.suimz.open.chatgptweb.java.bean.resp.R; import org.springframework.http.ResponseEntity; import org.springframework.http.converter.HttpMessageNotReadableException; import org.springframework.validation.BindException; import org.springframework.validation.FieldError; import org.springframework.web.bind.MethodArgumentNotValidException; import org.springframework.web.bind.annotation.ExceptionHandler; import org.springframework.web.bind.annotation.RestControllerAdvice; import org.springframework.web.method.annotation.MethodArgumentTypeMismatchException; import java.util.ArrayList; import java.util.HashMap; import java.util.List; import java.util.Map; /** * Handle Exceptions * * @author https://github.com/suimz */ @RestControllerAdvice public class AdviceException { @ExceptionHandler(BizException.class) public ResponseEntity<R> map(BizException e) { return ResponseEntity.status(e.getHttpStatus()).body(R.error(e.getMessage())); } @ExceptionHandler(UnauthorizedBizException.class) public R map(UnauthorizedBizException e) { return R.builder() .status("Unauthorized") .message(e.getMessage()) .build(); } @ExceptionHandler(HttpMessageNotReadableException.class) public R map(HttpMessageNotReadableException e) { return R.error("bad request"); } @ExceptionHandler(BindException.class) public R map(BindException e) { List<FieldError> fieldErrors = e.getBindingResult().getFieldErrors(); Map<String, Object> error = this.getValidError(fieldErrors);
return R.error(error.get("errorMsg").toString());
} @ExceptionHandler(MethodArgumentNotValidException.class) public R map(MethodArgumentNotValidException e) { List<FieldError> fieldErrors = e.getBindingResult().getFieldErrors(); Map<String, Object> error = this.getValidError(fieldErrors); return R.error(error.get("errorMsg").toString()); } private Map<String, Object> getValidError(List<FieldError> fieldErrors) { Map<String, Object> map = new HashMap<String, Object>(16); List<String> errorList = new ArrayList<String>(); StringBuffer errorMsg = new StringBuffer(); for (FieldError error : fieldErrors) { errorList.add(error.getDefaultMessage()); errorMsg.append(error.getDefaultMessage()); // first break; } map.put("errorList", errorList); map.put("errorMsg", errorMsg); return map; } @ExceptionHandler(MethodArgumentTypeMismatchException.class) public R map(MethodArgumentTypeMismatchException e) { return R.error(e.getMessage()); } @ExceptionHandler(Exception.class) public R map(Exception e) { return R.error(); } }
src/main/java/com/suimz/open/chatgptweb/java/core/exception/AdviceException.java
suimz-chatgpt-web-java-37a4c63
[ { "filename": "src/main/java/com/suimz/open/chatgptweb/java/controller/AppController.java", "retrieved_chunk": " thrEx = (BizException) e;\n } else if (e instanceof OpenAiHttpException) {\n OpenAiHttpException exception = (OpenAiHttpException) e;\n thrEx = new ApiRequestErrorBizException(exception.statusCode, exception.getMessage());\n } else if (e.getCause() instanceof SocketTimeoutException) {\n thrEx = new ApiRequestErrorBizException(0);\n } else if (e.getCause() instanceof SocketException || e.getCause() instanceof IOException) {\n thrEx = new ApiRequestErrorBizException(-1);\n } else {\n thrEx = new ApiRequestErrorBizException();", "score": 0.8062316179275513 }, { "filename": "src/main/java/com/suimz/open/chatgptweb/java/bean/resp/R.java", "retrieved_chunk": " public static <T> R<T> ok(T data) {\n return R.ok(null, data);\n }\n public static <T> R<T> ok(String message, T data) {\n return (R<T>) R.builder().status(\"Success\").message(message).data(data).build();\n }\n public static R error() {\n return R.error(\"服务异常 | server exception\");\n }\n public static R error(String error) {", "score": 0.8056484460830688 }, { "filename": "src/main/java/com/suimz/open/chatgptweb/java/bean/po/ReverseResponseBodyCallback.java", "retrieved_chunk": " ResponseBody errorBody = response.errorBody();\n if (errorBody == null) {\n throw e;\n } else {\n OpenAiError error = mapper.readValue(\n errorBody.string(),\n OpenAiError.class\n );\n throw new OpenAiHttpException(error, e, e.code());\n }", "score": 0.8019691109657288 }, { "filename": "src/main/java/com/suimz/open/chatgptweb/java/core/exception/ApiRequestErrorBizException.java", "retrieved_chunk": "package com.suimz.open.chatgptweb.java.core.exception;\nimport java.util.HashMap;\nimport java.util.Map;\npublic class ApiRequestErrorBizException extends BizException {\n public static final Map<Integer, String> OPENAI_HTTP_ERROR;\n static {\n OPENAI_HTTP_ERROR = new HashMap<>();\n OPENAI_HTTP_ERROR.put(-1, \"[OpenAI] 发送请求失败 | Failed to send request\");\n OPENAI_HTTP_ERROR.put(0, \"[OpenAI] 请求超时 | Request timeout\");\n OPENAI_HTTP_ERROR.put(401, \"[OpenAI] 提供错误的API密钥 | Incorrect API key provided\");", "score": 0.7960975766181946 }, { "filename": "src/main/java/com/suimz/open/chatgptweb/java/core/exception/ApiRequestErrorBizException.java", "retrieved_chunk": " OPENAI_HTTP_ERROR.put(403, \"[OpenAI] 服务器拒绝访问,请稍后再试 | Server refused to access, please try again later\");\n OPENAI_HTTP_ERROR.put(429, \"[OpenAI] 访问速率受限,请稍后再试 | Access rate is limited, please try again later\");\n OPENAI_HTTP_ERROR.put(500, \"[OpenAI] 服务器繁忙,请稍后再试 | Internal Server Error\");\n OPENAI_HTTP_ERROR.put(502, \"[OpenAI] 错误的网关 | Bad Gateway\");\n OPENAI_HTTP_ERROR.put(503, \"[OpenAI] 服务器繁忙,请稍后再试 | Server is busy, please try again later\");\n OPENAI_HTTP_ERROR.put(504, \"[OpenAI] 网关超时 | Gateway Time-out\");\n }\n public ApiRequestErrorBizException() {\n this(500);\n }", "score": 0.7855531573295593 } ]
java
return R.error(error.get("errorMsg").toString());
package com.suimz.open.chatgptweb.java.core.exception; import com.suimz.open.chatgptweb.java.bean.resp.R; import org.springframework.http.ResponseEntity; import org.springframework.http.converter.HttpMessageNotReadableException; import org.springframework.validation.BindException; import org.springframework.validation.FieldError; import org.springframework.web.bind.MethodArgumentNotValidException; import org.springframework.web.bind.annotation.ExceptionHandler; import org.springframework.web.bind.annotation.RestControllerAdvice; import org.springframework.web.method.annotation.MethodArgumentTypeMismatchException; import java.util.ArrayList; import java.util.HashMap; import java.util.List; import java.util.Map; /** * Handle Exceptions * * @author https://github.com/suimz */ @RestControllerAdvice public class AdviceException { @ExceptionHandler(BizException.class) public ResponseEntity<R> map(BizException e) { return ResponseEntity.status(e.getHttpStatus()).body(R.error(e.getMessage())); } @ExceptionHandler(UnauthorizedBizException.class) public R map(UnauthorizedBizException e) { return R.builder() .status("Unauthorized") .message(e.getMessage()) .build(); } @ExceptionHandler(HttpMessageNotReadableException.class) public R map(HttpMessageNotReadableException e) { return R.error("bad request"); } @ExceptionHandler(BindException.class) public R map(BindException e) { List<FieldError> fieldErrors = e.getBindingResult().getFieldErrors(); Map<String, Object> error = this.getValidError(fieldErrors); return R.error(error.get("errorMsg").toString()); } @ExceptionHandler(MethodArgumentNotValidException.class) public R map(MethodArgumentNotValidException e) { List<FieldError> fieldErrors = e.getBindingResult().getFieldErrors(); Map<String, Object> error = this.getValidError(fieldErrors); return R.error(error.get("errorMsg").toString()); } private Map<String, Object> getValidError(List<FieldError> fieldErrors) { Map<String, Object> map = new HashMap<String, Object>(16); List<String> errorList = new ArrayList<String>(); StringBuffer errorMsg = new StringBuffer(); for (FieldError error : fieldErrors) { errorList.add(error.getDefaultMessage()); errorMsg.append(error.getDefaultMessage()); // first break; } map.put("errorList", errorList); map.put("errorMsg", errorMsg); return map; } @ExceptionHandler(MethodArgumentTypeMismatchException.class) public R map(MethodArgumentTypeMismatchException e) { return R.error(e.getMessage()); } @ExceptionHandler(Exception.class) public R map(Exception e) { return
R.error();
} }
src/main/java/com/suimz/open/chatgptweb/java/core/exception/AdviceException.java
suimz-chatgpt-web-java-37a4c63
[ { "filename": "src/main/java/com/suimz/open/chatgptweb/java/bean/resp/R.java", "retrieved_chunk": " public static <T> R<T> ok(T data) {\n return R.ok(null, data);\n }\n public static <T> R<T> ok(String message, T data) {\n return (R<T>) R.builder().status(\"Success\").message(message).data(data).build();\n }\n public static R error() {\n return R.error(\"服务异常 | server exception\");\n }\n public static R error(String error) {", "score": 0.8082683086395264 }, { "filename": "src/main/java/com/suimz/open/chatgptweb/java/bean/resp/R.java", "retrieved_chunk": "public class R<T> {\n private String status;\n private String message;\n private T data;\n public static R ok() {\n return R.ok(null);\n }\n public static R ok(String message) {\n return R.ok(null, null);\n }", "score": 0.7858605980873108 }, { "filename": "src/main/java/com/suimz/open/chatgptweb/java/controller/AppController.java", "retrieved_chunk": " thrEx = (BizException) e;\n } else if (e instanceof OpenAiHttpException) {\n OpenAiHttpException exception = (OpenAiHttpException) e;\n thrEx = new ApiRequestErrorBizException(exception.statusCode, exception.getMessage());\n } else if (e.getCause() instanceof SocketTimeoutException) {\n thrEx = new ApiRequestErrorBizException(0);\n } else if (e.getCause() instanceof SocketException || e.getCause() instanceof IOException) {\n thrEx = new ApiRequestErrorBizException(-1);\n } else {\n thrEx = new ApiRequestErrorBizException();", "score": 0.7803984880447388 }, { "filename": "src/main/java/com/suimz/open/chatgptweb/java/bean/po/ReverseResponseBodyCallback.java", "retrieved_chunk": " ResponseBody errorBody = response.errorBody();\n if (errorBody == null) {\n throw e;\n } else {\n OpenAiError error = mapper.readValue(\n errorBody.string(),\n OpenAiError.class\n );\n throw new OpenAiHttpException(error, e, e.code());\n }", "score": 0.772114098072052 }, { "filename": "src/main/java/com/suimz/open/chatgptweb/java/core/exception/ApiRequestErrorBizException.java", "retrieved_chunk": "package com.suimz.open.chatgptweb.java.core.exception;\nimport java.util.HashMap;\nimport java.util.Map;\npublic class ApiRequestErrorBizException extends BizException {\n public static final Map<Integer, String> OPENAI_HTTP_ERROR;\n static {\n OPENAI_HTTP_ERROR = new HashMap<>();\n OPENAI_HTTP_ERROR.put(-1, \"[OpenAI] 发送请求失败 | Failed to send request\");\n OPENAI_HTTP_ERROR.put(0, \"[OpenAI] 请求超时 | Request timeout\");\n OPENAI_HTTP_ERROR.put(401, \"[OpenAI] 提供错误的API密钥 | Incorrect API key provided\");", "score": 0.7711237668991089 } ]
java
R.error();
package com.suimz.open.chatgptweb.java.controller; import cn.hutool.core.util.ObjectUtil; import cn.hutool.core.util.StrUtil; import com.suimz.open.chatgptweb.java.bean.constant.ApiRunMode; import com.suimz.open.chatgptweb.java.bean.req.AuthVerifyReq; import com.suimz.open.chatgptweb.java.bean.req.ChatProcessReq; import com.suimz.open.chatgptweb.java.bean.resp.ConfigResp; import com.suimz.open.chatgptweb.java.bean.resp.R; import com.suimz.open.chatgptweb.java.bean.resp.SessionResp; import com.suimz.open.chatgptweb.java.core.component.ratelimiter.ApiRateLimiter; import com.suimz.open.chatgptweb.java.core.exception.ApiRequestErrorBizException; import com.suimz.open.chatgptweb.java.core.exception.BizException; import com.suimz.open.chatgptweb.java.core.properties.AppProperties; import com.suimz.open.chatgptweb.java.service.OpenAiApiService; import com.suimz.open.chatgptweb.java.service.OpenAiReverseService; import com.suimz.open.chatgptweb.java.util.ObjUtil; import com.theokanning.openai.OpenAiHttpException; import javax.annotation.Resource; import lombok.extern.slf4j.Slf4j; import org.springframework.scheduling.concurrent.ThreadPoolTaskExecutor; import org.springframework.validation.annotation.Validated; import org.springframework.web.bind.annotation.PostMapping; import org.springframework.web.bind.annotation.RequestBody; import org.springframework.web.bind.annotation.RequestMapping; import org.springframework.web.bind.annotation.RestController; import org.springframework.web.servlet.mvc.method.annotation.SseEmitter; import java.io.IOException; import java.net.SocketException; import java.net.SocketTimeoutException; /** * Server API Controller * * @author https://github.com/suimz */ @RestController @RequestMapping("/api") @Slf4j public class AppController { @Resource private AppProperties appProperties; @Resource private ThreadPoolTaskExecutor asyncTaskExecutor; @Resource private OpenAiApiService openAiApiService; @Resource private OpenAiReverseService openAiReverseService; @PostMapping("/session") public R<SessionResp> session() { return R.ok( SessionResp.builder() .auth(StrUtil.isNotBlank(appProperties.getAuthSecretKey())) .model(ApiRunMode.get(appProperties).getName()) .build() ); } @PostMapping("/config") public R<ConfigResp> config() { String socksProxy; if (appProperties.getSocksProxy() != null && ObjectUtil.isAllNotEmpty(appProperties.getSocksProxy().getHost(), appProperties.getSocksProxy().getPort())) { socksProxy = StrUtil.format("{}:{}", appProperties.getSocksProxy().getHost(), appProperties.getSocksProxy().getPort()); } else { socksProxy = "-"; } String httpProxy; if (appProperties.getHttpProxy() != null && ObjectUtil.isAllNotEmpty(appProperties.getHttpProxy().getHost(), appProperties.getHttpProxy().getPort())) { httpProxy = StrUtil.format("{}:{}", appProperties.getHttpProxy().getHost(), appProperties.getHttpProxy().getPort()); } else { httpProxy = "-"; } Double balance = null; ApiRunMode apiRunMode = ApiRunMode.get(appProperties); if (apiRunMode == ApiRunMode.API) { balance = openAiApiService.queryBalance(); } return R.ok( ConfigResp.builder() .apiModel(apiRunMode.getName()) .timeoutMs(appProperties.getApiTimeoutMs()) .httpsProxy(httpProxy) .socksProxy(socksProxy) .reverseProxy(ObjUtil.getNotBlankValSequential("-", appProperties.getOpenaiReverseApiProxyUrl())) .balance(ObjUtil.getNotNullValSequential("-", balance)) .build() ); } @PostMapping("/verify") public R<SessionResp> authVerify(@RequestBody @Validated AuthVerifyReq req) { if (!StrUtil.equals(appProperties.getAuthSecretKey(), req.getToken())) { throw new BizException("Secret key is invalid"); } return
R.ok("Verify successfully");
} @ApiRateLimiter @PostMapping("/chat-process") public SseEmitter chatProcess(@RequestBody @Validated ChatProcessReq req) { SseEmitter sseEmitter = new SseEmitter(appProperties.getApiTimeoutMs()); asyncTaskExecutor.execute(() -> { try { switch (ApiRunMode.get(appProperties)) { case API: openAiApiService.streamChat(sseEmitter, req); break; case REVERSE: openAiReverseService.streamChat(sseEmitter, req); break; } } catch (Throwable e) { log.error(e.getMessage(), e); BizException thrEx; if (e instanceof BizException) { thrEx = (BizException) e; } else if (e instanceof OpenAiHttpException) { OpenAiHttpException exception = (OpenAiHttpException) e; thrEx = new ApiRequestErrorBizException(exception.statusCode, exception.getMessage()); } else if (e.getCause() instanceof SocketTimeoutException) { thrEx = new ApiRequestErrorBizException(0); } else if (e.getCause() instanceof SocketException || e.getCause() instanceof IOException) { thrEx = new ApiRequestErrorBizException(-1); } else { thrEx = new ApiRequestErrorBizException(); } sseEmitter.completeWithError(thrEx); } }); return sseEmitter; } }
src/main/java/com/suimz/open/chatgptweb/java/controller/AppController.java
suimz-chatgpt-web-java-37a4c63
[ { "filename": "src/main/java/com/suimz/open/chatgptweb/java/bean/req/AuthVerifyReq.java", "retrieved_chunk": "package com.suimz.open.chatgptweb.java.bean.req;\nimport javax.validation.constraints.NotBlank;\nimport lombok.Data;\n/**\n * Request Params - verify\n *\n * @author https://github.com/suimz\n */\n@Data\npublic class AuthVerifyReq {", "score": 0.840381920337677 }, { "filename": "src/main/java/com/suimz/open/chatgptweb/java/core/interceptor/AuthInterceptor.java", "retrieved_chunk": " if (StrUtil.isNotBlank(appProperties.getAuthSecretKey())) {\n String token = getToken(request);\n if (!StrUtil.equals(appProperties.getAuthSecretKey(), token)) {\n throw new UnauthorizedBizException();\n }\n }\n return true;\n }\n public String getToken(HttpServletRequest request) {\n String bearer = request.getHeader(\"Authorization\");", "score": 0.8224048018455505 }, { "filename": "src/main/java/com/suimz/open/chatgptweb/java/service/OpenAiReverseService.java", "retrieved_chunk": " }\n public void checkService() {\n if (StrUtil.isBlank(appProperties.getOpenaiAccessToken()) || StrUtil.isBlank(appProperties.getOpenaiReverseApiProxyUrl())) {\n throw new ReverseServiceNotInitializedBizException();\n }\n }\n @Override\n public void streamChat(SseEmitter sseEmitter, ChatProcessReq req) {\n this.checkService();\n String authHeader = \"Bearer \" + appProperties.getOpenaiAccessToken();", "score": 0.8087218999862671 }, { "filename": "src/main/java/com/suimz/open/chatgptweb/java/bean/resp/SessionResp.java", "retrieved_chunk": "@Getter\npublic class SessionResp {\n private boolean auth;\n /**\n * @see ApiRunMode\n */\n private String model;\n}", "score": 0.8058079481124878 }, { "filename": "src/main/java/com/suimz/open/chatgptweb/java/service/OpenAiApiService.java", "retrieved_chunk": " }\n public void checkService() {\n if (StrUtil.isBlank(appProperties.getOpenaiApiKey())) {\n throw new ApiServiceNotInitializedBizException();\n }\n }\n @Override\n public void streamChat(SseEmitter sseEmitter, ChatProcessReq req) {\n this.checkService();\n LocalDateTime startTime = LocalDateTime.now();", "score": 0.7930646538734436 } ]
java
R.ok("Verify successfully");
/* * Licensed to the Apache Software Foundation (ASF) under one or more * contributor license agreements. See the NOTICE file distributed with * this work for additional information regarding copyright ownership. * The ASF licenses this file to You under the Apache License, Version 2.0 * (the "License"); you may not use this file except in compliance with * the License. You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ package org.tallison.cc.index.extractor; import java.io.IOException; import java.util.Optional; import org.slf4j.Logger; import org.slf4j.LoggerFactory; import org.tallison.cc.index.AbstractRecordProcessor; import org.tallison.cc.index.CCIndexReaderCounter; import org.tallison.cc.index.CCIndexRecord; import org.apache.tika.exception.TikaConfigException; import org.apache.tika.utils.StringUtils; public class CCFileExtractorRecordProcessor extends AbstractRecordProcessor { private static Logger LOGGER = LoggerFactory.getLogger(CCFileExtractorRecordProcessor.class); private static Logger TRUNCATED_URLS_LOGGER = LoggerFactory.getLogger("truncated-urls"); private static Logger TRUNCATED_URLS_FULL_LOGGER = LoggerFactory.getLogger("truncated-urls-full"); private final ExtractorConfig fetcherConfig; private final CCIndexReaderCounter counter; private final FileFromCCWarcExtractor fileFromCCWarcFetcher; private long reportEvery = 100000; public CCFileExtractorRecordProcessor(ExtractorConfig fetcherConfig, CCIndexReaderCounter counter) throws TikaConfigException, IOException { this.fetcherConfig = fetcherConfig; this.counter = counter; this.fileFromCCWarcFetcher = new FileFromCCWarcExtractor(fetcherConfig, counter); //completely arbitrary if (fetcherConfig.getNumThreads() > 10) { reportEvery = 1000000; } } @Override public boolean process(String json) throws IOException, InterruptedException { //System.out.println("JSON: " + json); long totalRead = counter.getRecordsRead().incrementAndGet(); if (totalRead % reportEvery == 0) { LOGGER.info("processed: {}", counter); } if (fetcherConfig.getMaxRecords() > -1 && totalRead >= fetcherConfig.getMaxRecords()) { LOGGER.info("hit max read"); return false; } //check for hit max //return false; Optional<CCIndexRecord> record = CCIndexRecord.parseRecord(json); if (record.isEmpty()) { //problem already logged return true; } CCIndexRecord r = record.get(); if (!fetcherConfig.getRecordSelector().select(r)) { return true; } //if truncated, count appropriately and test for limits if (!StringUtils.isBlank(r.getTruncated())) { long truncated = counter.getTruncated().incrementAndGet(); if (fetcherConfig.getMaxFilesTruncated() > -1 && truncated >= fetcherConfig.getMaxFilesTruncated()) { LOGGER.info("hit max truncated files"); return false; } } if (fetcherConfig.isExtractTruncated() || StringUtils.isBlank(r.getTruncated())) { long extracted = counter.getFilesExtracted().incrementAndGet(); if (fetcherConfig.getMaxFilesExtracted() > -1 && extracted >= fetcherConfig.getMaxFilesExtracted()) { LOGGER.info("hit max extracted files"); return false; } if (fetcherConfig.isDryRun()) { LOGGER.info("dry run, but would have extracted {}", r); return true; } fetchBytes(r); return true; } else { String url = r.getUrl(); TRUNCATED_URLS_LOGGER.info("", url); //url,mime,mime_detected,warc_file,warc_offset,warc_length,truncated TRUNCATED_URLS_FULL_LOGGER.info("", url, r.getNormalizedMime(), r.getNormalizedMimeDetected(), r.getFilename(),
r.getOffset(), r.getLength(), r.getTruncated());
return true; } } private void fetchBytes(CCIndexRecord r) throws InterruptedException { fileFromCCWarcFetcher.fetchToPath(r); } @Override public void close() throws IOException { } }
src/main/java/org/tallison/cc/index/extractor/CCFileExtractorRecordProcessor.java
tballison-commoncrawl-fetcher-lite-2c106d1
[ { "filename": "src/main/java/org/tallison/cc/index/extractor/CCMimeCounter.java", "retrieved_chunk": " if (!fetcherConfig.getRecordSelector().select(r)) {\n return true;\n }\n increment(totalCounts, r.getNormalizedMimeDetected());\n if (!StringUtils.isBlank(r.getTruncated())) {\n long truncated = counter.getTruncated().incrementAndGet();\n if (fetcherConfig.getMaxFilesTruncated() > -1 &&\n truncated >= fetcherConfig.getMaxFilesTruncated()) {\n LOGGER.info(\"hit max truncated files\");\n return false;", "score": 0.8770785331726074 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/CCMimeCounter.java", "retrieved_chunk": " }\n increment(truncatedCounts, r.getNormalizedMimeDetected());\n return true;\n }\n return true;\n }\n private void increment(Map<String, MutableLong> m, String k) {\n MutableLong cnt = m.get(k);\n if (cnt == null) {\n cnt = new MutableLong(1);", "score": 0.8307200074195862 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/FileFromCCWarcExtractor.java", "retrieved_chunk": " //new ObjectArray ?\n //url,mime_detected,warc_file,warc_offset,warc_length,sha256,length,path\n EXTRACTED_LOGGER.info(\"\", ccIndexRecord.getUrl(),\n ccIndexRecord.getNormalizedMime(),\n ccIndexRecord.getNormalizedMimeDetected(),\n ccIndexRecord.getFilename(),\n ccIndexRecord.getOffset(), ccIndexRecord.getLength(),\n targetDigest, length,\n targetPath);\n }", "score": 0.8265665173530579 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/FileFromCCWarcExtractor.java", "retrieved_chunk": " String targetPath) {\n if (extractTruncated) {\n EXTRACTED_ALL_LOGGER.info(\"\", ccIndexRecord.getUrl(),\n ccIndexRecord.getNormalizedMime(),\n ccIndexRecord.getNormalizedMimeDetected(),\n ccIndexRecord.getFilename(),\n ccIndexRecord.getOffset(), ccIndexRecord.getLength(),\n ccIndexRecord.getTruncated(), targetDigest, length,\n targetPath);\n } else {", "score": 0.8260364532470703 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/FileFromCCWarcExtractor.java", "retrieved_chunk": " FetchEmitTuple t = new FetchEmitTuple(record.getFilename(),\n new FetchKey(\"\", record.getFilename(), record.getOffset(),\n record.getOffset() + record.getLength() - 1), new EmitKey());\n byte[] warcRecordGZBytes;\n try {\n warcRecordGZBytes = fetchWarcBytes(t);\n } catch (TikaException | IOException e) {\n LOGGER.warn(\"couldn't get bytes from cc's warc \" + t, e);\n return;\n }", "score": 0.8048960566520691 } ]
java
r.getOffset(), r.getLength(), r.getTruncated());
/* * Licensed to the Apache Software Foundation (ASF) under one or more * contributor license agreements. See the NOTICE file distributed with * this work for additional information regarding copyright ownership. * The ASF licenses this file to You under the Apache License, Version 2.0 * (the "License"); you may not use this file except in compliance with * the License. You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ package org.tallison.cc.index.selector; import java.util.HashMap; import java.util.List; import java.util.Map; import com.fasterxml.jackson.annotation.JsonProperty; import org.slf4j.Logger; import org.slf4j.LoggerFactory; import org.tallison.cc.index.CCIndexRecord; public class RecordSelector { private static Logger LOGGER = LoggerFactory.getLogger(RecordSelector.class); public static RecordSelector ACCEPT_ALL_RECORDS = new AcceptAllRecords(); @JsonProperty Map<String, List<SelectorClause>> must = new HashMap<>(); @JsonProperty Map<String, List<SelectorClause>> must_not = new HashMap<>(); @JsonProperty Map<String, List<SelectorClause>> should = new HashMap<>(); public boolean select(CCIndexRecord record) { for (Map.Entry<String, List<SelectorClause>> e : must_not.entrySet()) { String val = getStringValue(e.getKey(), record); if (val == null) { LOGGER.warn("Value is null for '{}' in the must not clause", e.getKey()); continue; } for (SelectorClause clause : e.getValue()) { if (clause.select(val)) { return false; } } } for (Map.Entry<String, List<SelectorClause>> e : must.entrySet()) { String val = getStringValue(e.getKey(), record); if (val == null) { LOGGER.warn("Value is null for '{}' in the must clause. Record not selected.", e.getKey()); return false; } for (SelectorClause clause : e.getValue()) { if (
!clause.select(val)) {
return false; } } } if (should.size() == 0) { return true; } for (Map.Entry<String, List<SelectorClause>> e : should.entrySet()) { String val = getStringValue(e.getKey(), record); if (val == null) { LOGGER.warn("Value is null for '{}' in the should clause. Record not selected", e.getKey()); continue; } for (SelectorClause clause : e.getValue()) { if (clause.select(val)) { return true; } } } return false; } private String getStringValue(String key, CCIndexRecord record) { switch (key) { case "mime_detected": return record.getMimeDetected(); case "truncated": return record.getTruncated(); case "mime": return record.getMime(); case "status": return Integer.toString(record.getStatus()); case "url": return record.getUrl(); case "host": return record.getHost(); case "digest": return record.getDigest(); default: throw new IllegalArgumentException("Don't yet support key " + key); } } private static class AcceptAllRecords extends RecordSelector { @Override public boolean select(CCIndexRecord record) { return true; } } }
src/main/java/org/tallison/cc/index/selector/RecordSelector.java
tballison-commoncrawl-fetcher-lite-2c106d1
[ { "filename": "src/main/java/org/tallison/cc/index/extractor/CCFileExtractorRecordProcessor.java", "retrieved_chunk": " //problem already logged\n return true;\n }\n CCIndexRecord r = record.get();\n if (!fetcherConfig.getRecordSelector().select(r)) {\n return true;\n }\n //if truncated, count appropriately and test for limits\n if (!StringUtils.isBlank(r.getTruncated())) {\n long truncated = counter.getTruncated().incrementAndGet();", "score": 0.8035719394683838 }, { "filename": "src/main/java/org/tallison/cc/index/selector/SelectorClause.java", "retrieved_chunk": "@JsonSubTypes({@JsonSubTypes.Type(MatchSelector.class), @JsonSubTypes.Type(RegexSelector.class)})\npublic interface SelectorClause {\n boolean select(String val);\n}", "score": 0.7914845943450928 }, { "filename": "src/main/java/org/tallison/cc/index/selector/MatchSelector.java", "retrieved_chunk": " this.match = match;\n this.caseSensitive = caseSensitive == null ? DEFAULT_CASE_SENSITIVE : caseSensitive;\n }\n @Override\n public boolean select(String val) {\n if (caseSensitive) {\n if (match.equals(val)) {\n return true;\n }\n } else {", "score": 0.789813220500946 }, { "filename": "src/main/java/org/tallison/cc/index/selector/RegexSelector.java", "retrieved_chunk": " public boolean select(String val) {\n Matcher m = pattern.matcher(val);\n if (m.find()) {\n return sampler.select(val);\n }\n return false;\n }\n}", "score": 0.789339542388916 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/CCFileExtractor.java", "retrieved_chunk": " fetchEmitTuple.getFetchKey().getFetchKey());\n while (line != null) {\n LOGGER.trace(\"about to add a line\");\n if (StringUtils.isBlank(line)) {\n line = reader.readLine();\n continue;\n }\n try {\n boolean shouldContinue = recordProcessor.process(line);\n if (!shouldContinue) {", "score": 0.7813300490379333 } ]
java
!clause.select(val)) {
package com.suimz.open.chatgptweb.java.core.component.ratelimiter; import com.suimz.open.chatgptweb.java.core.exception.ChatApiRequestTooManyBizException; import com.suimz.open.chatgptweb.java.core.properties.AppProperties; import com.suimz.open.chatgptweb.java.util.SpringUtil; import io.github.resilience4j.ratelimiter.RateLimiter; import io.github.resilience4j.ratelimiter.RateLimiterConfig; import io.github.resilience4j.ratelimiter.RateLimiterRegistry; import org.aspectj.lang.JoinPoint; import org.aspectj.lang.annotation.Aspect; import org.aspectj.lang.annotation.Before; import org.aspectj.lang.reflect.MethodSignature; import org.springframework.stereotype.Component; import java.lang.reflect.Method; import java.time.Duration; /** * API request rate restriction - Handle * * @author https://github.com/suimz */ @Aspect @Component public class ApiRateLimiterAspect { private RateLimiterRegistry rateLimiterRegistry = null; public ApiRateLimiterAspect(AppProperties appProperties) { Integer period = appProperties.getMaxRequestPerHour(); if (period != null && period > 0) { this.rateLimiterRegistry = RateLimiterRegistry.of( RateLimiterConfig.custom() .limitForPeriod(period) // Maximum number of requests .limitRefreshPeriod(Duration.ofHours(1)) // 1 hour .timeoutDuration(Duration.ofMillis(1)) .build() ); } } @Before("@annotation(apiRateLimiter)") public void doBefore(JoinPoint point, ApiRateLimiter apiRateLimiter) { if (this.rateLimiterRegistry == null) return; RateLimiter rateLimiter = rateLimiterRegistry.rateLimiter(getCombineKey(point)); if (!rateLimiter.acquirePermission()) throw new ChatApiRequestTooManyBizException(); } public String getCombineKey(JoinPoint point) {
StringBuilder sb = new StringBuilder(SpringUtil.getClientIp()).append("-");
MethodSignature signature = (MethodSignature) point.getSignature(); Method method = signature.getMethod(); Class<?> targetClass = method.getDeclaringClass(); sb.append(targetClass.getName()).append("-").append(method.getName()); return sb.toString(); } }
src/main/java/com/suimz/open/chatgptweb/java/core/component/ratelimiter/ApiRateLimiterAspect.java
suimz-chatgpt-web-java-37a4c63
[ { "filename": "src/main/java/com/suimz/open/chatgptweb/java/controller/AppController.java", "retrieved_chunk": "package com.suimz.open.chatgptweb.java.controller;\nimport cn.hutool.core.util.ObjectUtil;\nimport cn.hutool.core.util.StrUtil;\nimport com.suimz.open.chatgptweb.java.bean.constant.ApiRunMode;\nimport com.suimz.open.chatgptweb.java.bean.req.AuthVerifyReq;\nimport com.suimz.open.chatgptweb.java.bean.req.ChatProcessReq;\nimport com.suimz.open.chatgptweb.java.bean.resp.ConfigResp;\nimport com.suimz.open.chatgptweb.java.bean.resp.R;\nimport com.suimz.open.chatgptweb.java.bean.resp.SessionResp;\nimport com.suimz.open.chatgptweb.java.core.component.ratelimiter.ApiRateLimiter;", "score": 0.8064218163490295 }, { "filename": "src/main/java/com/suimz/open/chatgptweb/java/core/exception/ChatApiRequestTooManyBizException.java", "retrieved_chunk": "package com.suimz.open.chatgptweb.java.core.exception;\npublic class ChatApiRequestTooManyBizException extends BizException {\n public ChatApiRequestTooManyBizException() {\n super(\"Too many request from this IP in 1 hour\");\n }\n}", "score": 0.8045974969863892 }, { "filename": "src/main/java/com/suimz/open/chatgptweb/java/core/component/ratelimiter/ApiRateLimiter.java", "retrieved_chunk": "public @interface ApiRateLimiter {\n}", "score": 0.8043479919433594 }, { "filename": "src/main/java/com/suimz/open/chatgptweb/java/controller/AppController.java", "retrieved_chunk": "import com.suimz.open.chatgptweb.java.core.exception.ApiRequestErrorBizException;\nimport com.suimz.open.chatgptweb.java.core.exception.BizException;\nimport com.suimz.open.chatgptweb.java.core.properties.AppProperties;\nimport com.suimz.open.chatgptweb.java.service.OpenAiApiService;\nimport com.suimz.open.chatgptweb.java.service.OpenAiReverseService;\nimport com.suimz.open.chatgptweb.java.util.ObjUtil;\nimport com.theokanning.openai.OpenAiHttpException;\nimport javax.annotation.Resource;\nimport lombok.extern.slf4j.Slf4j;\nimport org.springframework.scheduling.concurrent.ThreadPoolTaskExecutor;", "score": 0.789905309677124 }, { "filename": "src/main/java/com/suimz/open/chatgptweb/java/core/component/ratelimiter/ApiRateLimiter.java", "retrieved_chunk": "package com.suimz.open.chatgptweb.java.core.component.ratelimiter;\nimport java.lang.annotation.*;\n/**\n * API request rate restriction - Annotation\n *\n * @author https://github.com/suimz\n */\n@Target(ElementType.METHOD)\n@Retention(RetentionPolicy.RUNTIME)\n@Documented", "score": 0.7855537533760071 } ]
java
StringBuilder sb = new StringBuilder(SpringUtil.getClientIp()).append("-");
/* * Licensed to the Apache Software Foundation (ASF) under one or more * contributor license agreements. See the NOTICE file distributed with * this work for additional information regarding copyright ownership. * The ASF licenses this file to You under the Apache License, Version 2.0 * (the "License"); you may not use this file except in compliance with * the License. You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ package org.tallison.cc.index.extractor; import java.io.IOException; import java.util.Optional; import org.slf4j.Logger; import org.slf4j.LoggerFactory; import org.tallison.cc.index.AbstractRecordProcessor; import org.tallison.cc.index.CCIndexReaderCounter; import org.tallison.cc.index.CCIndexRecord; import org.apache.tika.exception.TikaConfigException; import org.apache.tika.utils.StringUtils; public class CCFileExtractorRecordProcessor extends AbstractRecordProcessor { private static Logger LOGGER = LoggerFactory.getLogger(CCFileExtractorRecordProcessor.class); private static Logger TRUNCATED_URLS_LOGGER = LoggerFactory.getLogger("truncated-urls"); private static Logger TRUNCATED_URLS_FULL_LOGGER = LoggerFactory.getLogger("truncated-urls-full"); private final ExtractorConfig fetcherConfig; private final CCIndexReaderCounter counter; private final FileFromCCWarcExtractor fileFromCCWarcFetcher; private long reportEvery = 100000; public CCFileExtractorRecordProcessor(ExtractorConfig fetcherConfig, CCIndexReaderCounter counter) throws TikaConfigException, IOException { this.fetcherConfig = fetcherConfig; this.counter = counter; this.fileFromCCWarcFetcher = new FileFromCCWarcExtractor(fetcherConfig, counter); //completely arbitrary if (fetcherConfig.getNumThreads() > 10) { reportEvery = 1000000; } } @Override public boolean process(String json) throws IOException, InterruptedException { //System.out.println("JSON: " + json); long totalRead = counter.getRecordsRead().incrementAndGet(); if (totalRead % reportEvery == 0) { LOGGER.info("processed: {}", counter); } if (fetcherConfig.getMaxRecords() > -1 && totalRead >= fetcherConfig.getMaxRecords()) { LOGGER.info("hit max read"); return false; } //check for hit max //return false; Optional<CCIndexRecord> record = CCIndexRecord.parseRecord(json); if (record.isEmpty()) { //problem already logged return true; } CCIndexRecord r = record.get(); if (!fetcherConfig.getRecordSelector().select(r)) { return true; } //if truncated, count appropriately and test for limits if (!StringUtils.isBlank(r.getTruncated())) { long truncated = counter.getTruncated().incrementAndGet(); if (fetcherConfig.getMaxFilesTruncated() > -1 && truncated >= fetcherConfig.getMaxFilesTruncated()) { LOGGER.info("hit max truncated files"); return false; } } if (fetcherConfig.isExtractTruncated() || StringUtils.isBlank(r.getTruncated())) { long extracted = counter.getFilesExtracted().incrementAndGet(); if (fetcherConfig.getMaxFilesExtracted() > -1 && extracted >= fetcherConfig.getMaxFilesExtracted()) { LOGGER.info("hit max extracted files"); return false; } if (fetcherConfig.isDryRun()) { LOGGER.info("dry run, but would have extracted {}", r); return true; } fetchBytes(r); return true; } else { String url = r.getUrl(); TRUNCATED_URLS_LOGGER.info("", url); //url,mime,mime_detected,warc_file,warc_offset,warc_length,truncated TRUNCATED_URLS_FULL_LOGGER.info("", url, r.getNormalizedMime(), r.getNormalizedMimeDetected(), r.getFilename(), r.getOffset(
), r.getLength(), r.getTruncated());
return true; } } private void fetchBytes(CCIndexRecord r) throws InterruptedException { fileFromCCWarcFetcher.fetchToPath(r); } @Override public void close() throws IOException { } }
src/main/java/org/tallison/cc/index/extractor/CCFileExtractorRecordProcessor.java
tballison-commoncrawl-fetcher-lite-2c106d1
[ { "filename": "src/main/java/org/tallison/cc/index/extractor/CCMimeCounter.java", "retrieved_chunk": " if (!fetcherConfig.getRecordSelector().select(r)) {\n return true;\n }\n increment(totalCounts, r.getNormalizedMimeDetected());\n if (!StringUtils.isBlank(r.getTruncated())) {\n long truncated = counter.getTruncated().incrementAndGet();\n if (fetcherConfig.getMaxFilesTruncated() > -1 &&\n truncated >= fetcherConfig.getMaxFilesTruncated()) {\n LOGGER.info(\"hit max truncated files\");\n return false;", "score": 0.8800023794174194 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/CCMimeCounter.java", "retrieved_chunk": " }\n increment(truncatedCounts, r.getNormalizedMimeDetected());\n return true;\n }\n return true;\n }\n private void increment(Map<String, MutableLong> m, String k) {\n MutableLong cnt = m.get(k);\n if (cnt == null) {\n cnt = new MutableLong(1);", "score": 0.8306905627250671 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/FileFromCCWarcExtractor.java", "retrieved_chunk": " //new ObjectArray ?\n //url,mime_detected,warc_file,warc_offset,warc_length,sha256,length,path\n EXTRACTED_LOGGER.info(\"\", ccIndexRecord.getUrl(),\n ccIndexRecord.getNormalizedMime(),\n ccIndexRecord.getNormalizedMimeDetected(),\n ccIndexRecord.getFilename(),\n ccIndexRecord.getOffset(), ccIndexRecord.getLength(),\n targetDigest, length,\n targetPath);\n }", "score": 0.8255743980407715 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/FileFromCCWarcExtractor.java", "retrieved_chunk": " String targetPath) {\n if (extractTruncated) {\n EXTRACTED_ALL_LOGGER.info(\"\", ccIndexRecord.getUrl(),\n ccIndexRecord.getNormalizedMime(),\n ccIndexRecord.getNormalizedMimeDetected(),\n ccIndexRecord.getFilename(),\n ccIndexRecord.getOffset(), ccIndexRecord.getLength(),\n ccIndexRecord.getTruncated(), targetDigest, length,\n targetPath);\n } else {", "score": 0.8247002363204956 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/FileFromCCWarcExtractor.java", "retrieved_chunk": " FetchEmitTuple t = new FetchEmitTuple(record.getFilename(),\n new FetchKey(\"\", record.getFilename(), record.getOffset(),\n record.getOffset() + record.getLength() - 1), new EmitKey());\n byte[] warcRecordGZBytes;\n try {\n warcRecordGZBytes = fetchWarcBytes(t);\n } catch (TikaException | IOException e) {\n LOGGER.warn(\"couldn't get bytes from cc's warc \" + t, e);\n return;\n }", "score": 0.8075100779533386 } ]
java
), r.getLength(), r.getTruncated());
package com.suimz.open.chatgptweb.java.bean.po; import com.fasterxml.jackson.databind.ObjectMapper; import com.theokanning.openai.OpenAiError; import com.theokanning.openai.OpenAiHttpException; import com.theokanning.openai.service.OpenAiService; import com.theokanning.openai.service.SSEFormatException; import io.reactivex.FlowableEmitter; import okhttp3.ResponseBody; import retrofit2.Call; import retrofit2.Callback; import retrofit2.HttpException; import retrofit2.Response; import java.io.BufferedReader; import java.io.IOException; import java.io.InputStream; import java.io.InputStreamReader; public class ReverseResponseBodyCallback implements Callback<ResponseBody> { private static final ObjectMapper mapper = OpenAiService.defaultObjectMapper(); private FlowableEmitter<ReverseSSE> emitter; private boolean emitDone; public ReverseResponseBodyCallback(FlowableEmitter<ReverseSSE> emitter, boolean emitDone) { this.emitter = emitter; this.emitDone = emitDone; } @Override public void onResponse(Call<ResponseBody> call, Response<ResponseBody> response) { BufferedReader reader = null; try { if (!response.isSuccessful()) { HttpException e = new HttpException(response); ResponseBody errorBody = response.errorBody(); if (errorBody == null) { throw e; } else { OpenAiError error = mapper.readValue( errorBody.string(), OpenAiError.class ); throw new OpenAiHttpException(error, e, e.code()); } } InputStream in = response.body().byteStream(); reader = new BufferedReader(new InputStreamReader(in)); String line; ReverseSSE sse = null; while ((line = reader.readLine()) != null) { if (line.startsWith("data:")) { String data = line.substring(5).trim(); sse = new ReverseSSE(data); } else if (line.equals("") && sse != null) { if (
sse.isDone()) {
if (emitDone) { emitter.onNext(sse); } break; } emitter.onNext(sse); sse = null; } else { throw new SSEFormatException("Invalid sse format! " + line); } } emitter.onComplete(); } catch (Throwable t) { onFailure(call, t); } finally { if (reader != null) { try { reader.close(); } catch (IOException e) { // do nothing } } } } @Override public void onFailure(Call<ResponseBody> call, Throwable t) { emitter.onError(t); } }
src/main/java/com/suimz/open/chatgptweb/java/bean/po/ReverseResponseBodyCallback.java
suimz-chatgpt-web-java-37a4c63
[ { "filename": "src/main/java/com/suimz/open/chatgptweb/java/bean/po/ReverseSSE.java", "retrieved_chunk": "package com.suimz.open.chatgptweb.java.bean.po;\npublic class ReverseSSE {\n private static final String DONE_DATA = \"[DONE]\";\n private final String data;\n public ReverseSSE(String data){\n this.data = data;\n }\n public String getData(){\n return this.data;\n }", "score": 0.8318712711334229 }, { "filename": "src/main/java/com/suimz/open/chatgptweb/java/service/OpenAiReverseService.java", "retrieved_chunk": " ObjectNode body = buildSendMsgBody(req);\n String sendMsgId = body.findValues(\"messages\").get(0).findValue(\"id\").asText();\n Flowable.<ReverseSSE>create(emitter -> reverseApi.conversation(appProperties.getOpenaiReverseApiProxyUrl(), body, authHeader).enqueue(new ReverseResponseBodyCallback(emitter, false)), BackpressureStrategy.BUFFER)\n .map(sse -> okHttpObjectMapper.readValue(sse.getData(), ReverseChatChunk.class))\n .blockingForEach(chunk -> {\n try {\n if (StrUtil.isNotBlank(chunk.getError())) {\n log.debug(chunk.getError());\n sseEmitter.completeWithError(new BizException(chunk.getError()));\n }", "score": 0.7801274061203003 }, { "filename": "src/main/java/com/suimz/open/chatgptweb/java/bean/po/ReverseSSE.java", "retrieved_chunk": " public byte[] toBytes(){\n return String.format(\"data: %s\\n\\n\", this.data).getBytes();\n }\n public boolean isDone(){\n return DONE_DATA.equalsIgnoreCase(this.data);\n }\n}", "score": 0.772752583026886 }, { "filename": "src/main/java/com/suimz/open/chatgptweb/java/service/OpenAiReverseService.java", "retrieved_chunk": "package com.suimz.open.chatgptweb.java.service;\nimport cn.hutool.core.util.BooleanUtil;\nimport cn.hutool.core.util.IdUtil;\nimport cn.hutool.core.util.StrUtil;\nimport com.fasterxml.jackson.databind.node.ObjectNode;\nimport com.suimz.open.chatgptweb.java.bean.po.ReverseChatChunk;\nimport com.suimz.open.chatgptweb.java.bean.po.ReverseResponseBodyCallback;\nimport com.suimz.open.chatgptweb.java.bean.po.ReverseSSE;\nimport com.suimz.open.chatgptweb.java.bean.req.ChatProcessReq;\nimport com.suimz.open.chatgptweb.java.bean.resp.ChatProcessResp;", "score": 0.7710355520248413 }, { "filename": "src/main/java/com/suimz/open/chatgptweb/java/controller/AppController.java", "retrieved_chunk": " openAiApiService.streamChat(sseEmitter, req);\n break;\n case REVERSE:\n openAiReverseService.streamChat(sseEmitter, req);\n break;\n }\n } catch (Throwable e) {\n log.error(e.getMessage(), e);\n BizException thrEx;\n if (e instanceof BizException) {", "score": 0.7675004005432129 } ]
java
sse.isDone()) {
/* * Licensed to the Apache Software Foundation (ASF) under one or more * contributor license agreements. See the NOTICE file distributed with * this work for additional information regarding copyright ownership. * The ASF licenses this file to You under the Apache License, Version 2.0 * (the "License"); you may not use this file except in compliance with * the License. You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ package org.tallison.cc.index.extractor; import java.io.IOException; import java.util.Optional; import org.slf4j.Logger; import org.slf4j.LoggerFactory; import org.tallison.cc.index.AbstractRecordProcessor; import org.tallison.cc.index.CCIndexReaderCounter; import org.tallison.cc.index.CCIndexRecord; import org.apache.tika.exception.TikaConfigException; import org.apache.tika.utils.StringUtils; public class CCFileExtractorRecordProcessor extends AbstractRecordProcessor { private static Logger LOGGER = LoggerFactory.getLogger(CCFileExtractorRecordProcessor.class); private static Logger TRUNCATED_URLS_LOGGER = LoggerFactory.getLogger("truncated-urls"); private static Logger TRUNCATED_URLS_FULL_LOGGER = LoggerFactory.getLogger("truncated-urls-full"); private final ExtractorConfig fetcherConfig; private final CCIndexReaderCounter counter; private final FileFromCCWarcExtractor fileFromCCWarcFetcher; private long reportEvery = 100000; public CCFileExtractorRecordProcessor(ExtractorConfig fetcherConfig, CCIndexReaderCounter counter) throws TikaConfigException, IOException { this.fetcherConfig = fetcherConfig; this.counter = counter; this.fileFromCCWarcFetcher = new FileFromCCWarcExtractor(fetcherConfig, counter); //completely arbitrary if (fetcherConfig.getNumThreads() > 10) { reportEvery = 1000000; } } @Override public boolean process(String json) throws IOException, InterruptedException { //System.out.println("JSON: " + json); long totalRead = counter.getRecordsRead().incrementAndGet(); if (totalRead % reportEvery == 0) { LOGGER.info("processed: {}", counter); } if (fetcherConfig.getMaxRecords() > -1 && totalRead >= fetcherConfig.getMaxRecords()) { LOGGER.info("hit max read"); return false; } //check for hit max //return false; Optional<CCIndexRecord> record = CCIndexRecord.parseRecord(json); if (record.isEmpty()) { //problem already logged return true; } CCIndexRecord r = record.get(); if (!fetcherConfig.getRecordSelector().select(r)) { return true; } //if truncated, count appropriately and test for limits if (!StringUtils.isBlank(r.getTruncated())) { long truncated = counter.getTruncated().incrementAndGet(); if (fetcherConfig.getMaxFilesTruncated() > -1 && truncated >= fetcherConfig.getMaxFilesTruncated()) { LOGGER.info("hit max truncated files"); return false; } } if (fetcherConfig.isExtractTruncated() || StringUtils.isBlank(r.getTruncated())) { long
extracted = counter.getFilesExtracted().incrementAndGet();
if (fetcherConfig.getMaxFilesExtracted() > -1 && extracted >= fetcherConfig.getMaxFilesExtracted()) { LOGGER.info("hit max extracted files"); return false; } if (fetcherConfig.isDryRun()) { LOGGER.info("dry run, but would have extracted {}", r); return true; } fetchBytes(r); return true; } else { String url = r.getUrl(); TRUNCATED_URLS_LOGGER.info("", url); //url,mime,mime_detected,warc_file,warc_offset,warc_length,truncated TRUNCATED_URLS_FULL_LOGGER.info("", url, r.getNormalizedMime(), r.getNormalizedMimeDetected(), r.getFilename(), r.getOffset(), r.getLength(), r.getTruncated()); return true; } } private void fetchBytes(CCIndexRecord r) throws InterruptedException { fileFromCCWarcFetcher.fetchToPath(r); } @Override public void close() throws IOException { } }
src/main/java/org/tallison/cc/index/extractor/CCFileExtractorRecordProcessor.java
tballison-commoncrawl-fetcher-lite-2c106d1
[ { "filename": "src/main/java/org/tallison/cc/index/extractor/CCMimeCounter.java", "retrieved_chunk": " if (!fetcherConfig.getRecordSelector().select(r)) {\n return true;\n }\n increment(totalCounts, r.getNormalizedMimeDetected());\n if (!StringUtils.isBlank(r.getTruncated())) {\n long truncated = counter.getTruncated().incrementAndGet();\n if (fetcherConfig.getMaxFilesTruncated() > -1 &&\n truncated >= fetcherConfig.getMaxFilesTruncated()) {\n LOGGER.info(\"hit max truncated files\");\n return false;", "score": 0.941855788230896 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/CCMimeCounter.java", "retrieved_chunk": " return true;\n }\n }\n private static class DetectedMimeCounter extends AbstractRecordProcessor {\n private final ExtractorConfig fetcherConfig;\n private final CCIndexReaderCounter counter;\n private final Map<String, MutableLong> totalCounts = new HashMap<>();\n private final Map<String, MutableLong> truncatedCounts = new HashMap<>();\n public DetectedMimeCounter(ExtractorConfig fetcherConfig, CCIndexReaderCounter counter) {\n this.fetcherConfig = fetcherConfig;", "score": 0.8620455265045166 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/CCMimeCounter.java", "retrieved_chunk": " }\n increment(truncatedCounts, r.getNormalizedMimeDetected());\n return true;\n }\n return true;\n }\n private void increment(Map<String, MutableLong> m, String k) {\n MutableLong cnt = m.get(k);\n if (cnt == null) {\n cnt = new MutableLong(1);", "score": 0.8552988171577454 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/CCMimeCounter.java", "retrieved_chunk": " this.counter = counter;\n }\n @Override\n public boolean process(String json) throws IOException, InterruptedException {\n long totalRead = counter.getRecordsRead().incrementAndGet();\n if (totalRead % 1000000 == 0) {\n LOGGER.info(\"processed: {}\", counter);\n }\n if (fetcherConfig.getMaxRecords() > -1 && totalRead >= fetcherConfig.getMaxRecords()) {\n LOGGER.info(\"hit max read\");", "score": 0.8503425717353821 }, { "filename": "src/main/java/org/tallison/cc/index/CCIndexReaderCounter.java", "retrieved_chunk": " AtomicLong filesExtracted = new AtomicLong(0);\n AtomicLong truncated = new AtomicLong(0);\n AtomicLong emptyPayload = new AtomicLong(0);\n public AtomicLong getRecordsRead() {\n return recordsRead;\n }\n public AtomicLong getFilesExtracted() {\n return filesExtracted;\n }\n public AtomicLong getTruncated() {", "score": 0.830169677734375 } ]
java
extracted = counter.getFilesExtracted().incrementAndGet();
/* * Licensed to the Apache Software Foundation (ASF) under one or more * contributor license agreements. See the NOTICE file distributed with * this work for additional information regarding copyright ownership. * The ASF licenses this file to You under the Apache License, Version 2.0 * (the "License"); you may not use this file except in compliance with * the License. You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ package org.tallison.cc.index.extractor; import java.io.IOException; import java.util.Optional; import org.slf4j.Logger; import org.slf4j.LoggerFactory; import org.tallison.cc.index.AbstractRecordProcessor; import org.tallison.cc.index.CCIndexReaderCounter; import org.tallison.cc.index.CCIndexRecord; import org.apache.tika.exception.TikaConfigException; import org.apache.tika.utils.StringUtils; public class CCFileExtractorRecordProcessor extends AbstractRecordProcessor { private static Logger LOGGER = LoggerFactory.getLogger(CCFileExtractorRecordProcessor.class); private static Logger TRUNCATED_URLS_LOGGER = LoggerFactory.getLogger("truncated-urls"); private static Logger TRUNCATED_URLS_FULL_LOGGER = LoggerFactory.getLogger("truncated-urls-full"); private final ExtractorConfig fetcherConfig; private final CCIndexReaderCounter counter; private final FileFromCCWarcExtractor fileFromCCWarcFetcher; private long reportEvery = 100000; public CCFileExtractorRecordProcessor(ExtractorConfig fetcherConfig, CCIndexReaderCounter counter) throws TikaConfigException, IOException { this.fetcherConfig = fetcherConfig; this.counter = counter; this.fileFromCCWarcFetcher = new FileFromCCWarcExtractor(fetcherConfig, counter); //completely arbitrary if (fetcherConfig.getNumThreads() > 10) { reportEvery = 1000000; } } @Override public boolean process(String json) throws IOException, InterruptedException { //System.out.println("JSON: " + json); long totalRead = counter.getRecordsRead().incrementAndGet(); if (totalRead % reportEvery == 0) { LOGGER.info("processed: {}", counter); } if (fetcherConfig.getMaxRecords() > -1 && totalRead >= fetcherConfig.getMaxRecords()) { LOGGER.info("hit max read"); return false; } //check for hit max //return false; Optional<CCIndexRecord> record = CCIndexRecord.parseRecord(json); if (record.isEmpty()) { //problem already logged return true; } CCIndexRecord r = record.get(); if (!fetcherConfig.getRecordSelector().select(r)) { return true; } //if truncated, count appropriately and test for limits if (!StringUtils.isBlank(r.getTruncated())) { long truncated = counter.getTruncated().incrementAndGet(); if (fetcherConfig.getMaxFilesTruncated() > -1 && truncated >= fetcherConfig.getMaxFilesTruncated()) { LOGGER.info("hit max truncated files"); return false; } } if (fetcherConfig.isExtractTruncated() || StringUtils.isBlank(r.getTruncated())) { long extracted = counter.getFilesExtracted().incrementAndGet(); if (fetcherConfig.getMaxFilesExtracted() > -1 && extracted >= fetcherConfig.getMaxFilesExtracted()) { LOGGER.info("hit max extracted files"); return false; } if (fetcherConfig.isDryRun()) { LOGGER.info("dry run, but would have extracted {}", r); return true; } fetchBytes(r); return true; } else { String
url = r.getUrl();
TRUNCATED_URLS_LOGGER.info("", url); //url,mime,mime_detected,warc_file,warc_offset,warc_length,truncated TRUNCATED_URLS_FULL_LOGGER.info("", url, r.getNormalizedMime(), r.getNormalizedMimeDetected(), r.getFilename(), r.getOffset(), r.getLength(), r.getTruncated()); return true; } } private void fetchBytes(CCIndexRecord r) throws InterruptedException { fileFromCCWarcFetcher.fetchToPath(r); } @Override public void close() throws IOException { } }
src/main/java/org/tallison/cc/index/extractor/CCFileExtractorRecordProcessor.java
tballison-commoncrawl-fetcher-lite-2c106d1
[ { "filename": "src/main/java/org/tallison/cc/index/extractor/CCMimeCounter.java", "retrieved_chunk": " if (!fetcherConfig.getRecordSelector().select(r)) {\n return true;\n }\n increment(totalCounts, r.getNormalizedMimeDetected());\n if (!StringUtils.isBlank(r.getTruncated())) {\n long truncated = counter.getTruncated().incrementAndGet();\n if (fetcherConfig.getMaxFilesTruncated() > -1 &&\n truncated >= fetcherConfig.getMaxFilesTruncated()) {\n LOGGER.info(\"hit max truncated files\");\n return false;", "score": 0.8532355427742004 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/CCIndexFetcher.java", "retrieved_chunk": " return 1l;\n }\n fetch(t, fetcher, streamEmitter);\n }\n }\n private void fetch(FetchEmitTuple t, Fetcher fetcher, StreamEmitter streamEmitter) {\n LOGGER.info(\"about to download: \" + t.getFetchKey().getFetchKey());\n try (InputStream is = fetcher.fetch(t.getFetchKey().getFetchKey(), new Metadata())) {\n streamEmitter.emit(t.getFetchKey().getFetchKey(), is, new Metadata());\n LOGGER.info(\"successfully downloaded: \" + t.getFetchKey().getFetchKey());", "score": 0.8094460964202881 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/CCFileExtractor.java", "retrieved_chunk": " } catch (TikaException | IOException e) {\n LOGGER.error(\n \"failed while processing \" + fetchEmitTuple.getFetchKey().getFetchKey(), e);\n }\n long elapsed = System.currentTimeMillis() - start;\n LOGGER.info(\"finished processing index gz in ({}) ms: {}\",\n String.format(Locale.US, \"%,d\", elapsed),\n fetchEmitTuple.getFetchKey().getFetchKey());\n return true;\n }", "score": 0.8079376220703125 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/CCIndexFetcher.java", "retrieved_chunk": " } catch (TikaException | IOException e) {\n LOGGER.error(\"failed to copy \" + t.getFetchKey().getFetchKey(), e);\n }\n }\n }\n}", "score": 0.8019884824752808 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/FileFromCCWarcExtractor.java", "retrieved_chunk": " try (InputStream is = fetcher.fetch(k.getFetchKey(), k.getRangeStart(), k.getRangeEnd(),\n new Metadata())) {\n IOUtils.copy(is, bos);\n }\n return bos.toByteArray();\n }\n}", "score": 0.7947200536727905 } ]
java
url = r.getUrl();
/* * Licensed to the Apache Software Foundation (ASF) under one or more * contributor license agreements. See the NOTICE file distributed with * this work for additional information regarding copyright ownership. * The ASF licenses this file to You under the Apache License, Version 2.0 * (the "License"); you may not use this file except in compliance with * the License. You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ package org.tallison.cc.index.extractor; import java.io.IOException; import java.util.Optional; import org.slf4j.Logger; import org.slf4j.LoggerFactory; import org.tallison.cc.index.AbstractRecordProcessor; import org.tallison.cc.index.CCIndexReaderCounter; import org.tallison.cc.index.CCIndexRecord; import org.apache.tika.exception.TikaConfigException; import org.apache.tika.utils.StringUtils; public class CCFileExtractorRecordProcessor extends AbstractRecordProcessor { private static Logger LOGGER = LoggerFactory.getLogger(CCFileExtractorRecordProcessor.class); private static Logger TRUNCATED_URLS_LOGGER = LoggerFactory.getLogger("truncated-urls"); private static Logger TRUNCATED_URLS_FULL_LOGGER = LoggerFactory.getLogger("truncated-urls-full"); private final ExtractorConfig fetcherConfig; private final CCIndexReaderCounter counter; private final FileFromCCWarcExtractor fileFromCCWarcFetcher; private long reportEvery = 100000; public CCFileExtractorRecordProcessor(ExtractorConfig fetcherConfig, CCIndexReaderCounter counter) throws TikaConfigException, IOException { this.fetcherConfig = fetcherConfig; this.counter = counter; this.fileFromCCWarcFetcher = new FileFromCCWarcExtractor(fetcherConfig, counter); //completely arbitrary
if (fetcherConfig.getNumThreads() > 10) {
reportEvery = 1000000; } } @Override public boolean process(String json) throws IOException, InterruptedException { //System.out.println("JSON: " + json); long totalRead = counter.getRecordsRead().incrementAndGet(); if (totalRead % reportEvery == 0) { LOGGER.info("processed: {}", counter); } if (fetcherConfig.getMaxRecords() > -1 && totalRead >= fetcherConfig.getMaxRecords()) { LOGGER.info("hit max read"); return false; } //check for hit max //return false; Optional<CCIndexRecord> record = CCIndexRecord.parseRecord(json); if (record.isEmpty()) { //problem already logged return true; } CCIndexRecord r = record.get(); if (!fetcherConfig.getRecordSelector().select(r)) { return true; } //if truncated, count appropriately and test for limits if (!StringUtils.isBlank(r.getTruncated())) { long truncated = counter.getTruncated().incrementAndGet(); if (fetcherConfig.getMaxFilesTruncated() > -1 && truncated >= fetcherConfig.getMaxFilesTruncated()) { LOGGER.info("hit max truncated files"); return false; } } if (fetcherConfig.isExtractTruncated() || StringUtils.isBlank(r.getTruncated())) { long extracted = counter.getFilesExtracted().incrementAndGet(); if (fetcherConfig.getMaxFilesExtracted() > -1 && extracted >= fetcherConfig.getMaxFilesExtracted()) { LOGGER.info("hit max extracted files"); return false; } if (fetcherConfig.isDryRun()) { LOGGER.info("dry run, but would have extracted {}", r); return true; } fetchBytes(r); return true; } else { String url = r.getUrl(); TRUNCATED_URLS_LOGGER.info("", url); //url,mime,mime_detected,warc_file,warc_offset,warc_length,truncated TRUNCATED_URLS_FULL_LOGGER.info("", url, r.getNormalizedMime(), r.getNormalizedMimeDetected(), r.getFilename(), r.getOffset(), r.getLength(), r.getTruncated()); return true; } } private void fetchBytes(CCIndexRecord r) throws InterruptedException { fileFromCCWarcFetcher.fetchToPath(r); } @Override public void close() throws IOException { } }
src/main/java/org/tallison/cc/index/extractor/CCFileExtractorRecordProcessor.java
tballison-commoncrawl-fetcher-lite-2c106d1
[ { "filename": "src/main/java/org/tallison/cc/index/extractor/FileFromCCWarcExtractor.java", "retrieved_chunk": " LoggerFactory.getLogger(FileFromCCWarcExtractor.class);\n private static Logger EXTRACTED_LOGGER = LoggerFactory.getLogger(\"extracted-urls\");\n private static Logger EXTRACTED_ALL_LOGGER = LoggerFactory.getLogger(\"extracted-urls-all\");\n private final StreamEmitter emitter;\n private final TargetPathRewriter targetPathRewriter;\n private RangeFetcher fetcher;\n private final boolean extractTruncated;\n private Base32 base32 = new Base32();\n private final CCIndexReaderCounter ccIndexReaderCounter;\n public FileFromCCWarcExtractor(ExtractorConfig fetcherConfig,", "score": 0.8822425603866577 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/CCMimeCounter.java", "retrieved_chunk": " return true;\n }\n }\n private static class DetectedMimeCounter extends AbstractRecordProcessor {\n private final ExtractorConfig fetcherConfig;\n private final CCIndexReaderCounter counter;\n private final Map<String, MutableLong> totalCounts = new HashMap<>();\n private final Map<String, MutableLong> truncatedCounts = new HashMap<>();\n public DetectedMimeCounter(ExtractorConfig fetcherConfig, CCIndexReaderCounter counter) {\n this.fetcherConfig = fetcherConfig;", "score": 0.8785545825958252 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/FileFromCCWarcExtractor.java", "retrieved_chunk": " CCIndexReaderCounter ccIndexReaderCounter) throws TikaConfigException {\n this.emitter = fetcherConfig.newEmitter();\n this.fetcher = (RangeFetcher) fetcherConfig.newFetcher();\n this.targetPathRewriter = fetcherConfig.getTargetPathRewriter();\n this.extractTruncated = fetcherConfig.isExtractTruncated();\n this.ccIndexReaderCounter = ccIndexReaderCounter;\n }\n public void fetchToPath(CCIndexRecord record) throws InterruptedException {\n LOGGER.debug(\"going to fetch {} {}->{}\", record.getFilename(), record.getOffset(),\n record.getLength());", "score": 0.8733197450637817 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/CCMimeCounter.java", "retrieved_chunk": "import org.tallison.cc.index.AbstractRecordProcessor;\nimport org.tallison.cc.index.CCIndexReaderCounter;\nimport org.tallison.cc.index.CCIndexRecord;\nimport org.tallison.cc.index.IndexIterator;\nimport org.apache.tika.exception.TikaConfigException;\nimport org.apache.tika.exception.TikaException;\nimport org.apache.tika.io.TikaInputStream;\nimport org.apache.tika.metadata.Metadata;\nimport org.apache.tika.pipes.FetchEmitTuple;\nimport org.apache.tika.pipes.fetcher.Fetcher;", "score": 0.8662077188491821 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/CCMimeCounter.java", "retrieved_chunk": " private final Fetcher fetcher;\n IndexWorker(ExtractorConfig fetcherConfig, ArrayBlockingQueue<FetchEmitTuple> indexUrls,\n AbstractRecordProcessor recordProcessor) throws TikaException {\n this.indexUrls = indexUrls;\n this.recordProcessor = recordProcessor;\n this.fetcher = fetcherConfig.newFetcher();\n }\n @Override\n public Long call() throws Exception {\n boolean shouldContinue = true;", "score": 0.8644384145736694 } ]
java
if (fetcherConfig.getNumThreads() > 10) {
/* * Licensed to the Apache Software Foundation (ASF) under one or more * contributor license agreements. See the NOTICE file distributed with * this work for additional information regarding copyright ownership. * The ASF licenses this file to You under the Apache License, Version 2.0 * (the "License"); you may not use this file except in compliance with * the License. You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ package org.tallison.cc.index.extractor; import java.io.IOException; import java.util.Optional; import org.slf4j.Logger; import org.slf4j.LoggerFactory; import org.tallison.cc.index.AbstractRecordProcessor; import org.tallison.cc.index.CCIndexReaderCounter; import org.tallison.cc.index.CCIndexRecord; import org.apache.tika.exception.TikaConfigException; import org.apache.tika.utils.StringUtils; public class CCFileExtractorRecordProcessor extends AbstractRecordProcessor { private static Logger LOGGER = LoggerFactory.getLogger(CCFileExtractorRecordProcessor.class); private static Logger TRUNCATED_URLS_LOGGER = LoggerFactory.getLogger("truncated-urls"); private static Logger TRUNCATED_URLS_FULL_LOGGER = LoggerFactory.getLogger("truncated-urls-full"); private final ExtractorConfig fetcherConfig; private final CCIndexReaderCounter counter; private final FileFromCCWarcExtractor fileFromCCWarcFetcher; private long reportEvery = 100000; public CCFileExtractorRecordProcessor(ExtractorConfig fetcherConfig, CCIndexReaderCounter counter) throws TikaConfigException, IOException { this.fetcherConfig = fetcherConfig; this.counter = counter; this.fileFromCCWarcFetcher = new FileFromCCWarcExtractor(fetcherConfig, counter); //completely arbitrary if (fetcherConfig.getNumThreads() > 10) { reportEvery = 1000000; } } @Override public boolean process(String json) throws IOException, InterruptedException { //System.out.println("JSON: " + json); long totalRead = counter.getRecordsRead().incrementAndGet(); if (totalRead % reportEvery == 0) { LOGGER.info("processed: {}", counter); } if (fetcherConfig.getMaxRecords() > -1 && totalRead >= fetcherConfig.getMaxRecords()) { LOGGER.info("hit max read"); return false; } //check for hit max //return false; Optional<CCIndexRecord> record = CCIndexRecord.parseRecord(json); if (record.isEmpty()) { //problem already logged return true; } CCIndexRecord r = record.get(); if (!fetcherConfig.getRecordSelector().select(r)) { return true; } //if truncated, count appropriately and test for limits if (!StringUtils.isBlank(r.getTruncated())) { long truncated = counter.getTruncated().incrementAndGet(); if (fetcherConfig.getMaxFilesTruncated() > -1 && truncated >= fetcherConfig.getMaxFilesTruncated()) { LOGGER.info("hit max truncated files"); return false; } } if (fetcherConfig.isExtractTruncated() || StringUtils.isBlank(r.getTruncated())) { long extracted = counter.getFilesExtracted().incrementAndGet(); if (fetcherConfig.getMaxFilesExtracted() > -1 && extracted >= fetcherConfig.getMaxFilesExtracted()) { LOGGER.info("hit max extracted files"); return false; } if (fetcherConfig.isDryRun()) { LOGGER.info("dry run, but would have extracted {}", r); return true; } fetchBytes(r); return true; } else { String url = r.getUrl(); TRUNCATED_URLS_LOGGER.info("", url); //url,mime,mime_detected,warc_file,warc_offset,warc_length,truncated TRUNCATED_URLS_FULL_LOGGER.info("", url, r.getNormalizedMime(), r.
getNormalizedMimeDetected(), r.getFilename(), r.getOffset(), r.getLength(), r.getTruncated());
return true; } } private void fetchBytes(CCIndexRecord r) throws InterruptedException { fileFromCCWarcFetcher.fetchToPath(r); } @Override public void close() throws IOException { } }
src/main/java/org/tallison/cc/index/extractor/CCFileExtractorRecordProcessor.java
tballison-commoncrawl-fetcher-lite-2c106d1
[ { "filename": "src/main/java/org/tallison/cc/index/extractor/CCMimeCounter.java", "retrieved_chunk": " if (!fetcherConfig.getRecordSelector().select(r)) {\n return true;\n }\n increment(totalCounts, r.getNormalizedMimeDetected());\n if (!StringUtils.isBlank(r.getTruncated())) {\n long truncated = counter.getTruncated().incrementAndGet();\n if (fetcherConfig.getMaxFilesTruncated() > -1 &&\n truncated >= fetcherConfig.getMaxFilesTruncated()) {\n LOGGER.info(\"hit max truncated files\");\n return false;", "score": 0.87786865234375 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/CCMimeCounter.java", "retrieved_chunk": " }\n increment(truncatedCounts, r.getNormalizedMimeDetected());\n return true;\n }\n return true;\n }\n private void increment(Map<String, MutableLong> m, String k) {\n MutableLong cnt = m.get(k);\n if (cnt == null) {\n cnt = new MutableLong(1);", "score": 0.8284467458724976 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/FileFromCCWarcExtractor.java", "retrieved_chunk": " //new ObjectArray ?\n //url,mime_detected,warc_file,warc_offset,warc_length,sha256,length,path\n EXTRACTED_LOGGER.info(\"\", ccIndexRecord.getUrl(),\n ccIndexRecord.getNormalizedMime(),\n ccIndexRecord.getNormalizedMimeDetected(),\n ccIndexRecord.getFilename(),\n ccIndexRecord.getOffset(), ccIndexRecord.getLength(),\n targetDigest, length,\n targetPath);\n }", "score": 0.824352502822876 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/FileFromCCWarcExtractor.java", "retrieved_chunk": " String targetPath) {\n if (extractTruncated) {\n EXTRACTED_ALL_LOGGER.info(\"\", ccIndexRecord.getUrl(),\n ccIndexRecord.getNormalizedMime(),\n ccIndexRecord.getNormalizedMimeDetected(),\n ccIndexRecord.getFilename(),\n ccIndexRecord.getOffset(), ccIndexRecord.getLength(),\n ccIndexRecord.getTruncated(), targetDigest, length,\n targetPath);\n } else {", "score": 0.8230549693107605 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/FileFromCCWarcExtractor.java", "retrieved_chunk": " FetchEmitTuple t = new FetchEmitTuple(record.getFilename(),\n new FetchKey(\"\", record.getFilename(), record.getOffset(),\n record.getOffset() + record.getLength() - 1), new EmitKey());\n byte[] warcRecordGZBytes;\n try {\n warcRecordGZBytes = fetchWarcBytes(t);\n } catch (TikaException | IOException e) {\n LOGGER.warn(\"couldn't get bytes from cc's warc \" + t, e);\n return;\n }", "score": 0.807903528213501 } ]
java
getNormalizedMimeDetected(), r.getFilename(), r.getOffset(), r.getLength(), r.getTruncated());
/* * Licensed to the Apache Software Foundation (ASF) under one or more * contributor license agreements. See the NOTICE file distributed with * this work for additional information regarding copyright ownership. * The ASF licenses this file to You under the Apache License, Version 2.0 * (the "License"); you may not use this file except in compliance with * the License. You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ package org.tallison.cc.index.extractor; import java.io.IOException; import java.util.Optional; import org.slf4j.Logger; import org.slf4j.LoggerFactory; import org.tallison.cc.index.AbstractRecordProcessor; import org.tallison.cc.index.CCIndexReaderCounter; import org.tallison.cc.index.CCIndexRecord; import org.apache.tika.exception.TikaConfigException; import org.apache.tika.utils.StringUtils; public class CCFileExtractorRecordProcessor extends AbstractRecordProcessor { private static Logger LOGGER = LoggerFactory.getLogger(CCFileExtractorRecordProcessor.class); private static Logger TRUNCATED_URLS_LOGGER = LoggerFactory.getLogger("truncated-urls"); private static Logger TRUNCATED_URLS_FULL_LOGGER = LoggerFactory.getLogger("truncated-urls-full"); private final ExtractorConfig fetcherConfig; private final CCIndexReaderCounter counter; private final FileFromCCWarcExtractor fileFromCCWarcFetcher; private long reportEvery = 100000; public CCFileExtractorRecordProcessor(ExtractorConfig fetcherConfig, CCIndexReaderCounter counter) throws TikaConfigException, IOException { this.fetcherConfig = fetcherConfig; this.counter = counter; this.fileFromCCWarcFetcher = new FileFromCCWarcExtractor(fetcherConfig, counter); //completely arbitrary if (fetcherConfig.getNumThreads() > 10) { reportEvery = 1000000; } } @Override public boolean process(String json) throws IOException, InterruptedException { //System.out.println("JSON: " + json); long totalRead = counter.getRecordsRead().incrementAndGet(); if (totalRead % reportEvery == 0) { LOGGER.info("processed: {}", counter); } if (fetcherConfig.getMaxRecords() > -1 && totalRead >= fetcherConfig.getMaxRecords()) { LOGGER.info("hit max read"); return false; } //check for hit max //return false; Optional<CCIndexRecord> record = CCIndexRecord.parseRecord(json); if (record.isEmpty()) { //problem already logged return true; } CCIndexRecord r = record.get(); if (!fetcherConfig.getRecordSelector().select(r)) { return true; } //if truncated, count appropriately and test for limits if (!StringUtils.isBlank(r.getTruncated())) { long truncated = counter.getTruncated().incrementAndGet(); if
(fetcherConfig.getMaxFilesTruncated() > -1 && truncated >= fetcherConfig.getMaxFilesTruncated()) {
LOGGER.info("hit max truncated files"); return false; } } if (fetcherConfig.isExtractTruncated() || StringUtils.isBlank(r.getTruncated())) { long extracted = counter.getFilesExtracted().incrementAndGet(); if (fetcherConfig.getMaxFilesExtracted() > -1 && extracted >= fetcherConfig.getMaxFilesExtracted()) { LOGGER.info("hit max extracted files"); return false; } if (fetcherConfig.isDryRun()) { LOGGER.info("dry run, but would have extracted {}", r); return true; } fetchBytes(r); return true; } else { String url = r.getUrl(); TRUNCATED_URLS_LOGGER.info("", url); //url,mime,mime_detected,warc_file,warc_offset,warc_length,truncated TRUNCATED_URLS_FULL_LOGGER.info("", url, r.getNormalizedMime(), r.getNormalizedMimeDetected(), r.getFilename(), r.getOffset(), r.getLength(), r.getTruncated()); return true; } } private void fetchBytes(CCIndexRecord r) throws InterruptedException { fileFromCCWarcFetcher.fetchToPath(r); } @Override public void close() throws IOException { } }
src/main/java/org/tallison/cc/index/extractor/CCFileExtractorRecordProcessor.java
tballison-commoncrawl-fetcher-lite-2c106d1
[ { "filename": "src/main/java/org/tallison/cc/index/extractor/CCMimeCounter.java", "retrieved_chunk": " if (!fetcherConfig.getRecordSelector().select(r)) {\n return true;\n }\n increment(totalCounts, r.getNormalizedMimeDetected());\n if (!StringUtils.isBlank(r.getTruncated())) {\n long truncated = counter.getTruncated().incrementAndGet();\n if (fetcherConfig.getMaxFilesTruncated() > -1 &&\n truncated >= fetcherConfig.getMaxFilesTruncated()) {\n LOGGER.info(\"hit max truncated files\");\n return false;", "score": 0.9321501851081848 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/CCMimeCounter.java", "retrieved_chunk": " return true;\n }\n }\n private static class DetectedMimeCounter extends AbstractRecordProcessor {\n private final ExtractorConfig fetcherConfig;\n private final CCIndexReaderCounter counter;\n private final Map<String, MutableLong> totalCounts = new HashMap<>();\n private final Map<String, MutableLong> truncatedCounts = new HashMap<>();\n public DetectedMimeCounter(ExtractorConfig fetcherConfig, CCIndexReaderCounter counter) {\n this.fetcherConfig = fetcherConfig;", "score": 0.8605496287345886 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/CCMimeCounter.java", "retrieved_chunk": " return false;\n }\n //check for hit max\n //return false;\n Optional<CCIndexRecord> record = CCIndexRecord.parseRecord(json);\n if (record.isEmpty()) {\n //problem already logged\n return true;\n }\n CCIndexRecord r = record.get();", "score": 0.8374940156936646 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/FileFromCCWarcExtractor.java", "retrieved_chunk": " CCIndexReaderCounter ccIndexReaderCounter) throws TikaConfigException {\n this.emitter = fetcherConfig.newEmitter();\n this.fetcher = (RangeFetcher) fetcherConfig.newFetcher();\n this.targetPathRewriter = fetcherConfig.getTargetPathRewriter();\n this.extractTruncated = fetcherConfig.isExtractTruncated();\n this.ccIndexReaderCounter = ccIndexReaderCounter;\n }\n public void fetchToPath(CCIndexRecord record) throws InterruptedException {\n LOGGER.debug(\"going to fetch {} {}->{}\", record.getFilename(), record.getOffset(),\n record.getLength());", "score": 0.8342106938362122 }, { "filename": "src/main/java/org/tallison/cc/index/selector/AcceptAllRecords.java", "retrieved_chunk": " public boolean select(CCIndexRecord record) {\n return true;\n }\n}", "score": 0.8311145305633545 } ]
java
(fetcherConfig.getMaxFilesTruncated() > -1 && truncated >= fetcherConfig.getMaxFilesTruncated()) {
/* * Licensed to the Apache Software Foundation (ASF) under one or more * contributor license agreements. See the NOTICE file distributed with * this work for additional information regarding copyright ownership. * The ASF licenses this file to You under the Apache License, Version 2.0 * (the "License"); you may not use this file except in compliance with * the License. You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ package org.tallison.cc.index.extractor; import java.io.IOException; import java.util.Optional; import org.slf4j.Logger; import org.slf4j.LoggerFactory; import org.tallison.cc.index.AbstractRecordProcessor; import org.tallison.cc.index.CCIndexReaderCounter; import org.tallison.cc.index.CCIndexRecord; import org.apache.tika.exception.TikaConfigException; import org.apache.tika.utils.StringUtils; public class CCFileExtractorRecordProcessor extends AbstractRecordProcessor { private static Logger LOGGER = LoggerFactory.getLogger(CCFileExtractorRecordProcessor.class); private static Logger TRUNCATED_URLS_LOGGER = LoggerFactory.getLogger("truncated-urls"); private static Logger TRUNCATED_URLS_FULL_LOGGER = LoggerFactory.getLogger("truncated-urls-full"); private final ExtractorConfig fetcherConfig; private final CCIndexReaderCounter counter; private final FileFromCCWarcExtractor fileFromCCWarcFetcher; private long reportEvery = 100000; public CCFileExtractorRecordProcessor(ExtractorConfig fetcherConfig, CCIndexReaderCounter counter) throws TikaConfigException, IOException { this.fetcherConfig = fetcherConfig; this.counter = counter; this.fileFromCCWarcFetcher = new FileFromCCWarcExtractor(fetcherConfig, counter); //completely arbitrary if (fetcherConfig.getNumThreads() > 10) { reportEvery = 1000000; } } @Override public boolean process(String json) throws IOException, InterruptedException { //System.out.println("JSON: " + json); long totalRead = counter.getRecordsRead().incrementAndGet(); if (totalRead % reportEvery == 0) { LOGGER.info("processed: {}", counter); } if (fetcherConfig.getMaxRecords() > -1 && totalRead >= fetcherConfig.getMaxRecords()) { LOGGER.info("hit max read"); return false; } //check for hit max //return false; Optional<CCIndexRecord> record = CCIndexRecord.parseRecord(json); if (record.isEmpty()) { //problem already logged return true; } CCIndexRecord r = record.get(); if (!fetcherConfig.getRecordSelector().select(r)) { return true; } //if truncated, count appropriately and test for limits if (!StringUtils.isBlank(r.getTruncated())) { long truncated = counter.getTruncated().incrementAndGet(); if (fetcherConfig.getMaxFilesTruncated() > -1 && truncated >= fetcherConfig.getMaxFilesTruncated()) { LOGGER.info("hit max truncated files"); return false; } }
if (fetcherConfig.isExtractTruncated() || StringUtils.isBlank(r.getTruncated())) {
long extracted = counter.getFilesExtracted().incrementAndGet(); if (fetcherConfig.getMaxFilesExtracted() > -1 && extracted >= fetcherConfig.getMaxFilesExtracted()) { LOGGER.info("hit max extracted files"); return false; } if (fetcherConfig.isDryRun()) { LOGGER.info("dry run, but would have extracted {}", r); return true; } fetchBytes(r); return true; } else { String url = r.getUrl(); TRUNCATED_URLS_LOGGER.info("", url); //url,mime,mime_detected,warc_file,warc_offset,warc_length,truncated TRUNCATED_URLS_FULL_LOGGER.info("", url, r.getNormalizedMime(), r.getNormalizedMimeDetected(), r.getFilename(), r.getOffset(), r.getLength(), r.getTruncated()); return true; } } private void fetchBytes(CCIndexRecord r) throws InterruptedException { fileFromCCWarcFetcher.fetchToPath(r); } @Override public void close() throws IOException { } }
src/main/java/org/tallison/cc/index/extractor/CCFileExtractorRecordProcessor.java
tballison-commoncrawl-fetcher-lite-2c106d1
[ { "filename": "src/main/java/org/tallison/cc/index/extractor/CCMimeCounter.java", "retrieved_chunk": " if (!fetcherConfig.getRecordSelector().select(r)) {\n return true;\n }\n increment(totalCounts, r.getNormalizedMimeDetected());\n if (!StringUtils.isBlank(r.getTruncated())) {\n long truncated = counter.getTruncated().incrementAndGet();\n if (fetcherConfig.getMaxFilesTruncated() > -1 &&\n truncated >= fetcherConfig.getMaxFilesTruncated()) {\n LOGGER.info(\"hit max truncated files\");\n return false;", "score": 0.9430723190307617 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/CCMimeCounter.java", "retrieved_chunk": " }\n increment(truncatedCounts, r.getNormalizedMimeDetected());\n return true;\n }\n return true;\n }\n private void increment(Map<String, MutableLong> m, String k) {\n MutableLong cnt = m.get(k);\n if (cnt == null) {\n cnt = new MutableLong(1);", "score": 0.8513405323028564 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/CCMimeCounter.java", "retrieved_chunk": " return true;\n }\n }\n private static class DetectedMimeCounter extends AbstractRecordProcessor {\n private final ExtractorConfig fetcherConfig;\n private final CCIndexReaderCounter counter;\n private final Map<String, MutableLong> totalCounts = new HashMap<>();\n private final Map<String, MutableLong> truncatedCounts = new HashMap<>();\n public DetectedMimeCounter(ExtractorConfig fetcherConfig, CCIndexReaderCounter counter) {\n this.fetcherConfig = fetcherConfig;", "score": 0.8414039015769958 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/CCMimeCounter.java", "retrieved_chunk": " this.counter = counter;\n }\n @Override\n public boolean process(String json) throws IOException, InterruptedException {\n long totalRead = counter.getRecordsRead().incrementAndGet();\n if (totalRead % 1000000 == 0) {\n LOGGER.info(\"processed: {}\", counter);\n }\n if (fetcherConfig.getMaxRecords() > -1 && totalRead >= fetcherConfig.getMaxRecords()) {\n LOGGER.info(\"hit max read\");", "score": 0.8207046389579773 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/FileFromCCWarcExtractor.java", "retrieved_chunk": " String targetPath) {\n if (extractTruncated) {\n EXTRACTED_ALL_LOGGER.info(\"\", ccIndexRecord.getUrl(),\n ccIndexRecord.getNormalizedMime(),\n ccIndexRecord.getNormalizedMimeDetected(),\n ccIndexRecord.getFilename(),\n ccIndexRecord.getOffset(), ccIndexRecord.getLength(),\n ccIndexRecord.getTruncated(), targetDigest, length,\n targetPath);\n } else {", "score": 0.8077911734580994 } ]
java
if (fetcherConfig.isExtractTruncated() || StringUtils.isBlank(r.getTruncated())) {
/* * Licensed to the Apache Software Foundation (ASF) under one or more * contributor license agreements. See the NOTICE file distributed with * this work for additional information regarding copyright ownership. * The ASF licenses this file to You under the Apache License, Version 2.0 * (the "License"); you may not use this file except in compliance with * the License. You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ package org.tallison.cc.index.extractor; import java.io.IOException; import java.util.Optional; import org.slf4j.Logger; import org.slf4j.LoggerFactory; import org.tallison.cc.index.AbstractRecordProcessor; import org.tallison.cc.index.CCIndexReaderCounter; import org.tallison.cc.index.CCIndexRecord; import org.apache.tika.exception.TikaConfigException; import org.apache.tika.utils.StringUtils; public class CCFileExtractorRecordProcessor extends AbstractRecordProcessor { private static Logger LOGGER = LoggerFactory.getLogger(CCFileExtractorRecordProcessor.class); private static Logger TRUNCATED_URLS_LOGGER = LoggerFactory.getLogger("truncated-urls"); private static Logger TRUNCATED_URLS_FULL_LOGGER = LoggerFactory.getLogger("truncated-urls-full"); private final ExtractorConfig fetcherConfig; private final CCIndexReaderCounter counter; private final FileFromCCWarcExtractor fileFromCCWarcFetcher; private long reportEvery = 100000; public CCFileExtractorRecordProcessor(ExtractorConfig fetcherConfig, CCIndexReaderCounter counter) throws TikaConfigException, IOException { this.fetcherConfig = fetcherConfig; this.counter = counter; this.fileFromCCWarcFetcher = new FileFromCCWarcExtractor(fetcherConfig, counter); //completely arbitrary if (fetcherConfig.getNumThreads() > 10) { reportEvery = 1000000; } } @Override public boolean process(String json) throws IOException, InterruptedException { //System.out.println("JSON: " + json); long totalRead = counter.getRecordsRead().incrementAndGet(); if (totalRead % reportEvery == 0) { LOGGER.info("processed: {}", counter); } if (fetcherConfig.getMaxRecords() > -1 && totalRead >= fetcherConfig.getMaxRecords()) { LOGGER.info("hit max read"); return false; } //check for hit max //return false; Optional<CCIndexRecord> record = CCIndexRecord.parseRecord(json); if (record.isEmpty()) { //problem already logged return true; } CCIndexRecord r = record.get(); if (!fetcherConfig.getRecordSelector().select(r)) { return true; } //if truncated, count appropriately and test for limits if (!StringUtils.isBlank(r.getTruncated())) { long truncated = counter.getTruncated().incrementAndGet(); if (fetcherConfig.getMaxFilesTruncated() > -1 && truncated >= fetcherConfig.getMaxFilesTruncated()) { LOGGER.info("hit max truncated files"); return false; } } if (fetcherConfig.isExtractTruncated() || StringUtils.isBlank(r.getTruncated())) { long extracted = counter.getFilesExtracted().incrementAndGet(); if (fetcherConfig.getMaxFilesExtracted() > -1 && extracted >= fetcherConfig.getMaxFilesExtracted()) { LOGGER.info("hit max extracted files"); return false; }
if (fetcherConfig.isDryRun()) {
LOGGER.info("dry run, but would have extracted {}", r); return true; } fetchBytes(r); return true; } else { String url = r.getUrl(); TRUNCATED_URLS_LOGGER.info("", url); //url,mime,mime_detected,warc_file,warc_offset,warc_length,truncated TRUNCATED_URLS_FULL_LOGGER.info("", url, r.getNormalizedMime(), r.getNormalizedMimeDetected(), r.getFilename(), r.getOffset(), r.getLength(), r.getTruncated()); return true; } } private void fetchBytes(CCIndexRecord r) throws InterruptedException { fileFromCCWarcFetcher.fetchToPath(r); } @Override public void close() throws IOException { } }
src/main/java/org/tallison/cc/index/extractor/CCFileExtractorRecordProcessor.java
tballison-commoncrawl-fetcher-lite-2c106d1
[ { "filename": "src/main/java/org/tallison/cc/index/extractor/CCMimeCounter.java", "retrieved_chunk": " if (!fetcherConfig.getRecordSelector().select(r)) {\n return true;\n }\n increment(totalCounts, r.getNormalizedMimeDetected());\n if (!StringUtils.isBlank(r.getTruncated())) {\n long truncated = counter.getTruncated().incrementAndGet();\n if (fetcherConfig.getMaxFilesTruncated() > -1 &&\n truncated >= fetcherConfig.getMaxFilesTruncated()) {\n LOGGER.info(\"hit max truncated files\");\n return false;", "score": 0.932003378868103 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/CCMimeCounter.java", "retrieved_chunk": " return true;\n }\n }\n private static class DetectedMimeCounter extends AbstractRecordProcessor {\n private final ExtractorConfig fetcherConfig;\n private final CCIndexReaderCounter counter;\n private final Map<String, MutableLong> totalCounts = new HashMap<>();\n private final Map<String, MutableLong> truncatedCounts = new HashMap<>();\n public DetectedMimeCounter(ExtractorConfig fetcherConfig, CCIndexReaderCounter counter) {\n this.fetcherConfig = fetcherConfig;", "score": 0.8479404449462891 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/CCMimeCounter.java", "retrieved_chunk": " this.counter = counter;\n }\n @Override\n public boolean process(String json) throws IOException, InterruptedException {\n long totalRead = counter.getRecordsRead().incrementAndGet();\n if (totalRead % 1000000 == 0) {\n LOGGER.info(\"processed: {}\", counter);\n }\n if (fetcherConfig.getMaxRecords() > -1 && totalRead >= fetcherConfig.getMaxRecords()) {\n LOGGER.info(\"hit max read\");", "score": 0.8411847352981567 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/CCMimeCounter.java", "retrieved_chunk": " }\n increment(truncatedCounts, r.getNormalizedMimeDetected());\n return true;\n }\n return true;\n }\n private void increment(Map<String, MutableLong> m, String k) {\n MutableLong cnt = m.get(k);\n if (cnt == null) {\n cnt = new MutableLong(1);", "score": 0.8297069072723389 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/CCFileExtractor.java", "retrieved_chunk": " } catch (TikaException | IOException e) {\n LOGGER.error(\n \"failed while processing \" + fetchEmitTuple.getFetchKey().getFetchKey(), e);\n }\n long elapsed = System.currentTimeMillis() - start;\n LOGGER.info(\"finished processing index gz in ({}) ms: {}\",\n String.format(Locale.US, \"%,d\", elapsed),\n fetchEmitTuple.getFetchKey().getFetchKey());\n return true;\n }", "score": 0.8164857029914856 } ]
java
if (fetcherConfig.isDryRun()) {
/* * Licensed to the Apache Software Foundation (ASF) under one or more * contributor license agreements. See the NOTICE file distributed with * this work for additional information regarding copyright ownership. * The ASF licenses this file to You under the Apache License, Version 2.0 * (the "License"); you may not use this file except in compliance with * the License. You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ package org.tallison.cc.index.extractor; import java.io.BufferedInputStream; import java.io.BufferedReader; import java.io.File; import java.io.IOException; import java.io.InputStream; import java.io.InputStreamReader; import java.nio.charset.StandardCharsets; import java.util.Collections; import java.util.Locale; import java.util.concurrent.ArrayBlockingQueue; import java.util.concurrent.Callable; import java.util.concurrent.ExecutionException; import java.util.concurrent.ExecutorCompletionService; import java.util.concurrent.ExecutorService; import java.util.concurrent.Executors; import java.util.concurrent.Future; import java.util.concurrent.TimeUnit; import java.util.concurrent.TimeoutException; import java.util.zip.GZIPInputStream; import com.fasterxml.jackson.databind.ObjectMapper; import org.slf4j.Logger; import org.slf4j.LoggerFactory; import org.tallison.cc.index.AbstractRecordProcessor; import org.tallison.cc.index.CCIndexReaderCounter; import org.tallison.cc.index.IndexIterator; import org.apache.tika.exception.TikaConfigException; import org.apache.tika.exception.TikaException; import org.apache.tika.io.TikaInputStream; import org.apache.tika.metadata.Metadata; import org.apache.tika.pipes.FetchEmitTuple; import org.apache.tika.pipes.fetcher.Fetcher; import org.apache.tika.pipes.pipesiterator.CallablePipesIterator; import org.apache.tika.pipes.pipesiterator.PipesIterator; import org.apache.tika.utils.StringUtils; /** * This is a lighter class that doesn't rely on a database * to extract files from CC and log a list of truncated urls. */ public class CCFileExtractor { private static final Long INDEX_WORKER_ID = 1l; private static final Long INDEX_READER_ID = 2l; private static final Logger LOGGER = LoggerFactory.getLogger(CCFileExtractor.class); public static void main(String[] args) throws Exception { ExtractorConfig fetcherConfig = new ObjectMapper().readValue(new File(args[0]), ExtractorConfig.class); execute(fetcherConfig); } private static void execute(ExtractorConfig fetcherConfig) throws TikaException { ArrayBlockingQueue<FetchEmitTuple> indexPathsList = new ArrayBlockingQueue<>(1000); //IndexPathsReader reads a file containing a list of cc-index.paths files //and writes the literal gz files (cc-index/collections/CC-MAIN-2023-06/indexes/cdx-00000.gz) //to indexPathsList //IndexWorker reads a single index.gz file at a time and processes each record //It fetches non truncated files and logs truncated files int totalThreads = fetcherConfig.getNumThreads() + 1; ExecutorService executorService = Executors.newFixedThreadPool(totalThreads); ExecutorCompletionService<Long> executorCompletionService = new ExecutorCompletionService<>(executorService); IndexIterator indexIterator = fetcherConfig.getIndexIterator(); indexIterator.initialize(Collections.EMPTY_MAP); executorCompletionService.submit(new CallablePipesIterator(indexIterator, indexPathsList)); CCIndexReaderCounter counter = new CCIndexReaderCounter(); int finishedWorkers = 0; try { for (int i = 0; i < fetcherConfig.getNumThreads(); i++) { CCFileExtractorRecordProcessor processor = new CCFileExtractorRecordProcessor(fetcherConfig, counter); executorCompletionService.submit( new IndexWorker(fetcherConfig, indexPathsList, processor)); } while (finishedWorkers < fetcherConfig.getNumThreads()) { //blocking Future<Long> future = executorCompletionService.take(); if (future != null) { Long f = future.get(); LOGGER.debug("completed {}", f); if (f.equals(INDEX_WORKER_ID)) { finishedWorkers++; } else if (f.equals(INDEX_READER_ID)) { LOGGER.info("Index paths reader successfully completed"); } } } } catch (TikaConfigException | IOException e) { LOGGER.error("main loop exception", e); throw new RuntimeException(e); } catch (ExecutionException e) { LOGGER.error("main loop exception", e); throw new RuntimeException(e); } catch (InterruptedException e) { LOGGER.warn("main loop interrupted exception", e); throw new RuntimeException(e); } finally { executorService.shutdown(); executorService.shutdownNow(); } } private static class IndexWorker implements Callable<Long> { private final ArrayBlockingQueue<FetchEmitTuple> indexUrls; private final AbstractRecordProcessor recordProcessor; private final Fetcher indexFetcher; IndexWorker(ExtractorConfig fetcherConfig, ArrayBlockingQueue<FetchEmitTuple> indexUrls, AbstractRecordProcessor recordProcessor) throws TikaException { this.indexUrls = indexUrls; this.recordProcessor = recordProcessor; this
.indexFetcher = fetcherConfig.newIndexFetcher();
} @Override public Long call() throws Exception { boolean shouldContinue = true; while (shouldContinue) { FetchEmitTuple indexUrl = indexUrls.poll(120, TimeUnit.MINUTES); if (indexUrl == null) { throw new TimeoutException("waited 120 minutes for a new record"); } if (indexUrl == PipesIterator.COMPLETED_SEMAPHORE) { recordProcessor.close(); //can hang forever indexUrls.put(PipesIterator.COMPLETED_SEMAPHORE); return INDEX_WORKER_ID; } LOGGER.trace(indexUrl.toString()); shouldContinue = processFile(indexUrl, recordProcessor); } return INDEX_WORKER_ID; } private boolean processFile(FetchEmitTuple fetchEmitTuple, AbstractRecordProcessor recordProcessor) throws InterruptedException { long start = System.currentTimeMillis(); LOGGER.info("starting to fetch index gz: {}", fetchEmitTuple.getFetchKey().getFetchKey()); try (TikaInputStream tis = (TikaInputStream) indexFetcher.fetch( fetchEmitTuple.getFetchKey().getFetchKey(), new Metadata())) { try (InputStream is = new BufferedInputStream(new GZIPInputStream(tis))) { try (BufferedReader reader = new BufferedReader( new InputStreamReader(is, StandardCharsets.UTF_8))) { String line = reader.readLine(); int lines = 0; long elapsed = System.currentTimeMillis() - start; LOGGER.info("Finished fetching {} bytes in {} ms for index gz: {}", String.format(Locale.US, "%,d", tis.getLength()), String.format(Locale.US, "%,d", elapsed), fetchEmitTuple.getFetchKey().getFetchKey()); while (line != null) { LOGGER.trace("about to add a line"); if (StringUtils.isBlank(line)) { line = reader.readLine(); continue; } try { boolean shouldContinue = recordProcessor.process(line); if (!shouldContinue) { return shouldContinue; } } catch (IOException e) { LOGGER.warn("bad json: " + line); } lines++; line = reader.readLine(); } } } } catch (TikaException | IOException e) { LOGGER.error( "failed while processing " + fetchEmitTuple.getFetchKey().getFetchKey(), e); } long elapsed = System.currentTimeMillis() - start; LOGGER.info("finished processing index gz in ({}) ms: {}", String.format(Locale.US, "%,d", elapsed), fetchEmitTuple.getFetchKey().getFetchKey()); return true; } } }
src/main/java/org/tallison/cc/index/extractor/CCFileExtractor.java
tballison-commoncrawl-fetcher-lite-2c106d1
[ { "filename": "src/main/java/org/tallison/cc/index/extractor/CCMimeCounter.java", "retrieved_chunk": " private final Fetcher fetcher;\n IndexWorker(ExtractorConfig fetcherConfig, ArrayBlockingQueue<FetchEmitTuple> indexUrls,\n AbstractRecordProcessor recordProcessor) throws TikaException {\n this.indexUrls = indexUrls;\n this.recordProcessor = recordProcessor;\n this.fetcher = fetcherConfig.newFetcher();\n }\n @Override\n public Long call() throws Exception {\n boolean shouldContinue = true;", "score": 0.9646993279457092 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/CCIndexFetcher.java", "retrieved_chunk": " private static class IndexFetcher implements Callable<Long> {\n private final ExtractorConfig fetcherConfig;\n private final ArrayBlockingQueue<FetchEmitTuple> indexPathsList;\n public IndexFetcher(ExtractorConfig fetcherConfig,\n ArrayBlockingQueue<FetchEmitTuple> indexPathsList) {\n this.fetcherConfig = fetcherConfig;\n this.indexPathsList = indexPathsList;\n }\n @Override\n public Long call() throws Exception {", "score": 0.9281134605407715 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/CCMimeCounter.java", "retrieved_chunk": " if (cnt == null) {\n cnt = 0l;\n }\n cnt += e.getValue().getValue();\n to.put(e.getKey(), cnt);\n }\n }\n private static class IndexWorker implements Callable<Long> {\n private final ArrayBlockingQueue<FetchEmitTuple> indexUrls;\n private final AbstractRecordProcessor recordProcessor;", "score": 0.8995429873466492 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/CCMimeCounter.java", "retrieved_chunk": " }\n shouldContinue = processFile(indexUrl, recordProcessor);\n }\n return INDEX_WORKER_ID;\n }\n private boolean processFile(FetchEmitTuple fetchEmitTuple,\n AbstractRecordProcessor recordProcessor)\n throws InterruptedException {\n long start = System.currentTimeMillis();\n LOGGER.info(\"starting to fetch index gz: {}\",", "score": 0.8688459992408752 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/CCMimeCounter.java", "retrieved_chunk": " while (shouldContinue) {\n FetchEmitTuple indexUrl = indexUrls.poll(120, TimeUnit.MINUTES);\n if (indexUrl == null) {\n throw new TimeoutException(\"waited 120 minutes for a new record\");\n }\n if (indexUrl == PipesIterator.COMPLETED_SEMAPHORE) {\n recordProcessor.close();\n //can hang forever\n indexUrls.put(PipesIterator.COMPLETED_SEMAPHORE);\n return INDEX_WORKER_ID;", "score": 0.8687745332717896 } ]
java
.indexFetcher = fetcherConfig.newIndexFetcher();
/* * Licensed to the Apache Software Foundation (ASF) under one or more * contributor license agreements. See the NOTICE file distributed with * this work for additional information regarding copyright ownership. * The ASF licenses this file to You under the Apache License, Version 2.0 * (the "License"); you may not use this file except in compliance with * the License. You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ package org.tallison.cc.index.extractor; import java.io.ByteArrayInputStream; import java.io.ByteArrayOutputStream; import java.io.IOException; import java.io.InputStream; import java.nio.file.Files; import java.nio.file.Path; import java.nio.file.StandardCopyOption; import java.util.Optional; import java.util.zip.GZIPInputStream; import org.apache.commons.codec.binary.Base32; import org.apache.commons.codec.digest.DigestUtils; import org.apache.commons.io.IOUtils; import org.netpreserve.jwarc.MediaType; import org.netpreserve.jwarc.WarcPayload; import org.netpreserve.jwarc.WarcReader; import org.netpreserve.jwarc.WarcRecord; import org.netpreserve.jwarc.WarcResponse; import org.slf4j.Logger; import org.slf4j.LoggerFactory; import org.tallison.cc.index.CCIndexReaderCounter; import org.tallison.cc.index.CCIndexRecord; import org.tallison.cc.index.io.TargetPathRewriter; import org.apache.tika.exception.TikaConfigException; import org.apache.tika.exception.TikaException; import org.apache.tika.io.TikaInputStream; import org.apache.tika.metadata.Metadata; import org.apache.tika.pipes.FetchEmitTuple; import org.apache.tika.pipes.emitter.EmitKey; import org.apache.tika.pipes.emitter.StreamEmitter; import org.apache.tika.pipes.fetcher.FetchKey; import org.apache.tika.pipes.fetcher.RangeFetcher; public class FileFromCCWarcExtractor { private static Logger LOGGER = LoggerFactory.getLogger(FileFromCCWarcExtractor.class); private static Logger EXTRACTED_LOGGER = LoggerFactory.getLogger("extracted-urls"); private static Logger EXTRACTED_ALL_LOGGER = LoggerFactory.getLogger("extracted-urls-all"); private final StreamEmitter emitter; private final TargetPathRewriter targetPathRewriter; private RangeFetcher fetcher; private final boolean extractTruncated; private Base32 base32 = new Base32(); private final CCIndexReaderCounter ccIndexReaderCounter; public FileFromCCWarcExtractor(ExtractorConfig fetcherConfig, CCIndexReaderCounter ccIndexReaderCounter) throws TikaConfigException { this.emitter = fetcherConfig.newEmitter(); this.fetcher = (RangeFetcher) fetcherConfig.newFetcher(); this.targetPathRewriter = fetcherConfig.getTargetPathRewriter(); this.extractTruncated = fetcherConfig.isExtractTruncated(); this.ccIndexReaderCounter = ccIndexReaderCounter; } public void fetchToPath(CCIndexRecord record) throws InterruptedException { LOGGER.debug("going to fetch {} {}->{}", record.getFilename(), record.getOffset(), record.getLength()); FetchEmitTuple t = new FetchEmitTuple(record.getFilename(), new FetchKey("", record.getFilename(), record.getOffset(), record.getOffset() + record.getLength() - 1), new EmitKey()); byte[] warcRecordGZBytes; try { warcRecordGZBytes = fetchWarcBytes(t); } catch (TikaException | IOException e) { LOGGER.warn("couldn't get bytes from cc's warc " + t, e); return; } String id = record.getUrl(); try { parseWarc(id, record, warcRecordGZBytes); } catch (IOException e) { LOGGER.warn("problem parsing warc file", e); } } private void fetchPayload(String id, CCIndexRecord ccIndexRecord, WarcRecord record) throws IOException { if (!((record instanceof WarcResponse) && record.contentType().base().equals(MediaType.HTTP))) { return; } Optional<WarcPayload> payload = ((WarcResponse) record).payload(); if (!payload.isPresent()) { LOGGER.debug("no payload {}", id);
ccIndexReaderCounter.getEmptyPayload().incrementAndGet();
return; } if (payload.get().body().size() == 0) { LOGGER.debug("empty payload id={}", id); ccIndexReaderCounter.getEmptyPayload().incrementAndGet(); return; } Path tmp = Files.createTempFile("ccfile-fetcher-", ""); try { Files.copy(payload.get().body().stream(), tmp, StandardCopyOption.REPLACE_EXISTING); String targetDigest = null; String base32Sha1 = ""; try (InputStream is = Files.newInputStream(tmp)) { base32Sha1 = base32.encodeAsString(DigestUtils.sha1(is)); } catch (IOException e) { LOGGER.warn("IOException during digesting: " + tmp.toAbsolutePath()); return; } if (!base32Sha1.equals(ccIndexRecord.getDigest())) { LOGGER.warn("Bad digest for url={} ccindex={} sha1={}", id, ccIndexRecord.getDigest(), base32Sha1); } //TODO: make digest and encoding configurable try (InputStream is = Files.newInputStream(tmp)) { targetDigest = DigestUtils.sha256Hex(is); } catch (IOException e) { LOGGER.warn("IOException during digesting: " + tmp.toAbsolutePath()); return; } long length = -1; try { length = Files.size(tmp); } catch (IOException e) { LOGGER.warn("IOException during digesting: " + tmp.toAbsolutePath()); return; } String targetPath = targetPathRewriter.rewrite(targetDigest); Metadata metadata = new Metadata(); try (InputStream is = TikaInputStream.get(tmp, metadata)) { emitter.emit(targetPath, is, new Metadata()); logSuccess(ccIndexRecord, targetDigest, length, targetPath); } catch (IOException | TikaException e) { LOGGER.warn("problem writing id={}", id, e); } } finally { try { Files.delete(tmp); } catch (IOException e) { LOGGER.warn("can't delete " + tmp.toAbsolutePath(), e); } } } private void logSuccess(CCIndexRecord ccIndexRecord, String targetDigest, long length, String targetPath) { if (extractTruncated) { EXTRACTED_ALL_LOGGER.info("", ccIndexRecord.getUrl(), ccIndexRecord.getNormalizedMime(), ccIndexRecord.getNormalizedMimeDetected(), ccIndexRecord.getFilename(), ccIndexRecord.getOffset(), ccIndexRecord.getLength(), ccIndexRecord.getTruncated(), targetDigest, length, targetPath); } else { //new ObjectArray ? //url,mime_detected,warc_file,warc_offset,warc_length,sha256,length,path EXTRACTED_LOGGER.info("", ccIndexRecord.getUrl(), ccIndexRecord.getNormalizedMime(), ccIndexRecord.getNormalizedMimeDetected(), ccIndexRecord.getFilename(), ccIndexRecord.getOffset(), ccIndexRecord.getLength(), targetDigest, length, targetPath); } } private void parseWarc(String id, CCIndexRecord ccIndexRecord, byte[] warcRecordGZBytes) throws IOException { //need to leave initial inputstream open while parsing warcrecord //can't just parse record and return try (InputStream is = new GZIPInputStream(new ByteArrayInputStream(warcRecordGZBytes))) { try (WarcReader warcreader = new WarcReader(is)) { //should be a single warc per file //return the first for (WarcRecord warcRecord : warcreader) { fetchPayload(id, ccIndexRecord, warcRecord); return; } } } } private byte[] fetchWarcBytes(FetchEmitTuple t) throws TikaException, InterruptedException, IOException { ByteArrayOutputStream bos = new ByteArrayOutputStream(); FetchKey k = t.getFetchKey(); try (InputStream is = fetcher.fetch(k.getFetchKey(), k.getRangeStart(), k.getRangeEnd(), new Metadata())) { IOUtils.copy(is, bos); } return bos.toByteArray(); } }
src/main/java/org/tallison/cc/index/extractor/FileFromCCWarcExtractor.java
tballison-commoncrawl-fetcher-lite-2c106d1
[ { "filename": "src/main/java/org/tallison/cc/index/extractor/CCMimeCounter.java", "retrieved_chunk": "import org.tallison.cc.index.AbstractRecordProcessor;\nimport org.tallison.cc.index.CCIndexReaderCounter;\nimport org.tallison.cc.index.CCIndexRecord;\nimport org.tallison.cc.index.IndexIterator;\nimport org.apache.tika.exception.TikaConfigException;\nimport org.apache.tika.exception.TikaException;\nimport org.apache.tika.io.TikaInputStream;\nimport org.apache.tika.metadata.Metadata;\nimport org.apache.tika.pipes.FetchEmitTuple;\nimport org.apache.tika.pipes.fetcher.Fetcher;", "score": 0.8161294460296631 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/CCFileExtractorRecordProcessor.java", "retrieved_chunk": " String url = r.getUrl();\n TRUNCATED_URLS_LOGGER.info(\"\", url);\n //url,mime,mime_detected,warc_file,warc_offset,warc_length,truncated\n TRUNCATED_URLS_FULL_LOGGER.info(\"\", url,\n r.getNormalizedMime(), r.getNormalizedMimeDetected(), r.getFilename(),\n r.getOffset(), r.getLength(), r.getTruncated());\n return true;\n }\n }\n private void fetchBytes(CCIndexRecord r) throws InterruptedException {", "score": 0.8143547773361206 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/CCFileExtractorRecordProcessor.java", "retrieved_chunk": " //problem already logged\n return true;\n }\n CCIndexRecord r = record.get();\n if (!fetcherConfig.getRecordSelector().select(r)) {\n return true;\n }\n //if truncated, count appropriately and test for limits\n if (!StringUtils.isBlank(r.getTruncated())) {\n long truncated = counter.getTruncated().incrementAndGet();", "score": 0.8086895942687988 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/CCFileExtractorRecordProcessor.java", "retrieved_chunk": " LOGGER.info(\"processed: {}\", counter);\n }\n if (fetcherConfig.getMaxRecords() > -1 && totalRead >= fetcherConfig.getMaxRecords()) {\n LOGGER.info(\"hit max read\");\n return false;\n }\n //check for hit max\n //return false;\n Optional<CCIndexRecord> record = CCIndexRecord.parseRecord(json);\n if (record.isEmpty()) {", "score": 0.8075705766677856 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/CCFileExtractor.java", "retrieved_chunk": "import java.util.concurrent.ExecutorService;\nimport java.util.concurrent.Executors;\nimport java.util.concurrent.Future;\nimport java.util.concurrent.TimeUnit;\nimport java.util.concurrent.TimeoutException;\nimport java.util.zip.GZIPInputStream;\nimport com.fasterxml.jackson.databind.ObjectMapper;\nimport org.slf4j.Logger;\nimport org.slf4j.LoggerFactory;\nimport org.tallison.cc.index.AbstractRecordProcessor;", "score": 0.8028137683868408 } ]
java
ccIndexReaderCounter.getEmptyPayload().incrementAndGet();
/* * Licensed to the Apache Software Foundation (ASF) under one or more * contributor license agreements. See the NOTICE file distributed with * this work for additional information regarding copyright ownership. * The ASF licenses this file to You under the Apache License, Version 2.0 * (the "License"); you may not use this file except in compliance with * the License. You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ package org.tallison.cc.index.extractor; import java.io.BufferedInputStream; import java.io.BufferedReader; import java.io.File; import java.io.IOException; import java.io.InputStream; import java.io.InputStreamReader; import java.nio.charset.StandardCharsets; import java.util.Collections; import java.util.Locale; import java.util.concurrent.ArrayBlockingQueue; import java.util.concurrent.Callable; import java.util.concurrent.ExecutionException; import java.util.concurrent.ExecutorCompletionService; import java.util.concurrent.ExecutorService; import java.util.concurrent.Executors; import java.util.concurrent.Future; import java.util.concurrent.TimeUnit; import java.util.concurrent.TimeoutException; import java.util.zip.GZIPInputStream; import com.fasterxml.jackson.databind.ObjectMapper; import org.slf4j.Logger; import org.slf4j.LoggerFactory; import org.tallison.cc.index.AbstractRecordProcessor; import org.tallison.cc.index.CCIndexReaderCounter; import org.tallison.cc.index.IndexIterator; import org.apache.tika.exception.TikaConfigException; import org.apache.tika.exception.TikaException; import org.apache.tika.io.TikaInputStream; import org.apache.tika.metadata.Metadata; import org.apache.tika.pipes.FetchEmitTuple; import org.apache.tika.pipes.fetcher.Fetcher; import org.apache.tika.pipes.pipesiterator.CallablePipesIterator; import org.apache.tika.pipes.pipesiterator.PipesIterator; import org.apache.tika.utils.StringUtils; /** * This is a lighter class that doesn't rely on a database * to extract files from CC and log a list of truncated urls. */ public class CCFileExtractor { private static final Long INDEX_WORKER_ID = 1l; private static final Long INDEX_READER_ID = 2l; private static final Logger LOGGER = LoggerFactory.getLogger(CCFileExtractor.class); public static void main(String[] args) throws Exception { ExtractorConfig fetcherConfig = new ObjectMapper().readValue(new File(args[0]), ExtractorConfig.class); execute(fetcherConfig); } private static void execute(ExtractorConfig fetcherConfig) throws TikaException { ArrayBlockingQueue<FetchEmitTuple> indexPathsList = new ArrayBlockingQueue<>(1000); //IndexPathsReader reads a file containing a list of cc-index.paths files //and writes the literal gz files (cc-index/collections/CC-MAIN-2023-06/indexes/cdx-00000.gz) //to indexPathsList //IndexWorker reads a single index.gz file at a time and processes each record //It fetches non truncated files and logs truncated files int totalThreads = fetcherConfig.getNumThreads() + 1; ExecutorService executorService = Executors.newFixedThreadPool(totalThreads); ExecutorCompletionService<Long> executorCompletionService = new ExecutorCompletionService<>(executorService); IndexIterator indexIterator = fetcherConfig.getIndexIterator(); indexIterator.initialize(Collections.EMPTY_MAP); executorCompletionService.submit(new CallablePipesIterator(indexIterator, indexPathsList)); CCIndexReaderCounter counter = new CCIndexReaderCounter(); int finishedWorkers = 0; try { for (int i = 0; i < fetcherConfig.getNumThreads(); i++) { CCFileExtractorRecordProcessor processor = new CCFileExtractorRecordProcessor(fetcherConfig, counter); executorCompletionService.submit( new IndexWorker(fetcherConfig, indexPathsList, processor)); } while (
finishedWorkers < fetcherConfig.getNumThreads()) {
//blocking Future<Long> future = executorCompletionService.take(); if (future != null) { Long f = future.get(); LOGGER.debug("completed {}", f); if (f.equals(INDEX_WORKER_ID)) { finishedWorkers++; } else if (f.equals(INDEX_READER_ID)) { LOGGER.info("Index paths reader successfully completed"); } } } } catch (TikaConfigException | IOException e) { LOGGER.error("main loop exception", e); throw new RuntimeException(e); } catch (ExecutionException e) { LOGGER.error("main loop exception", e); throw new RuntimeException(e); } catch (InterruptedException e) { LOGGER.warn("main loop interrupted exception", e); throw new RuntimeException(e); } finally { executorService.shutdown(); executorService.shutdownNow(); } } private static class IndexWorker implements Callable<Long> { private final ArrayBlockingQueue<FetchEmitTuple> indexUrls; private final AbstractRecordProcessor recordProcessor; private final Fetcher indexFetcher; IndexWorker(ExtractorConfig fetcherConfig, ArrayBlockingQueue<FetchEmitTuple> indexUrls, AbstractRecordProcessor recordProcessor) throws TikaException { this.indexUrls = indexUrls; this.recordProcessor = recordProcessor; this.indexFetcher = fetcherConfig.newIndexFetcher(); } @Override public Long call() throws Exception { boolean shouldContinue = true; while (shouldContinue) { FetchEmitTuple indexUrl = indexUrls.poll(120, TimeUnit.MINUTES); if (indexUrl == null) { throw new TimeoutException("waited 120 minutes for a new record"); } if (indexUrl == PipesIterator.COMPLETED_SEMAPHORE) { recordProcessor.close(); //can hang forever indexUrls.put(PipesIterator.COMPLETED_SEMAPHORE); return INDEX_WORKER_ID; } LOGGER.trace(indexUrl.toString()); shouldContinue = processFile(indexUrl, recordProcessor); } return INDEX_WORKER_ID; } private boolean processFile(FetchEmitTuple fetchEmitTuple, AbstractRecordProcessor recordProcessor) throws InterruptedException { long start = System.currentTimeMillis(); LOGGER.info("starting to fetch index gz: {}", fetchEmitTuple.getFetchKey().getFetchKey()); try (TikaInputStream tis = (TikaInputStream) indexFetcher.fetch( fetchEmitTuple.getFetchKey().getFetchKey(), new Metadata())) { try (InputStream is = new BufferedInputStream(new GZIPInputStream(tis))) { try (BufferedReader reader = new BufferedReader( new InputStreamReader(is, StandardCharsets.UTF_8))) { String line = reader.readLine(); int lines = 0; long elapsed = System.currentTimeMillis() - start; LOGGER.info("Finished fetching {} bytes in {} ms for index gz: {}", String.format(Locale.US, "%,d", tis.getLength()), String.format(Locale.US, "%,d", elapsed), fetchEmitTuple.getFetchKey().getFetchKey()); while (line != null) { LOGGER.trace("about to add a line"); if (StringUtils.isBlank(line)) { line = reader.readLine(); continue; } try { boolean shouldContinue = recordProcessor.process(line); if (!shouldContinue) { return shouldContinue; } } catch (IOException e) { LOGGER.warn("bad json: " + line); } lines++; line = reader.readLine(); } } } } catch (TikaException | IOException e) { LOGGER.error( "failed while processing " + fetchEmitTuple.getFetchKey().getFetchKey(), e); } long elapsed = System.currentTimeMillis() - start; LOGGER.info("finished processing index gz in ({}) ms: {}", String.format(Locale.US, "%,d", elapsed), fetchEmitTuple.getFetchKey().getFetchKey()); return true; } } }
src/main/java/org/tallison/cc/index/extractor/CCFileExtractor.java
tballison-commoncrawl-fetcher-lite-2c106d1
[ { "filename": "src/main/java/org/tallison/cc/index/extractor/CCMimeCounter.java", "retrieved_chunk": " CCIndexReaderCounter counter = new CCIndexReaderCounter();\n int finishedWorkers = 0;\n List<DetectedMimeCounter> detectedMimeCounters = new ArrayList<>();\n try {\n for (int i = 0; i < fetcherConfig.getNumThreads(); i++) {\n DetectedMimeCounter processor = new DetectedMimeCounter(fetcherConfig, counter);\n detectedMimeCounters.add(processor);\n executorCompletionService.submit(\n new IndexWorker(fetcherConfig, indexPathsList, processor));\n }", "score": 0.9456884860992432 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/CCIndexFetcher.java", "retrieved_chunk": " ExecutorCompletionService<Long> executorCompletionService =\n new ExecutorCompletionService<>(executorService);\n IndexIterator indexIterator = fetcherConfig.getIndexIterator();\n indexIterator.initialize(Collections.EMPTY_MAP);\n executorCompletionService.submit(new CallablePipesIterator(indexIterator, indexPathsList));\n int finishedWorkers = 0;\n try {\n for (int i = 0; i < fetcherConfig.getNumThreads(); i++) {\n executorCompletionService.submit(new IndexFetcher(fetcherConfig, indexPathsList));\n }", "score": 0.9173789620399475 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/CCMimeCounter.java", "retrieved_chunk": " //to indexPathsList\n //IndexWorker reads a single index.gz file at a time and processes each record\n //It fetches non truncated files and logs truncated files\n int totalThreads = fetcherConfig.getNumThreads() + 1;\n ExecutorService executorService = Executors.newFixedThreadPool(totalThreads);\n ExecutorCompletionService<Long> executorCompletionService =\n new ExecutorCompletionService<>(executorService);\n IndexIterator indexIterator = fetcherConfig.getIndexIterator();\n indexIterator.initialize(Collections.EMPTY_MAP);\n executorCompletionService.submit(new CallablePipesIterator(indexIterator, indexPathsList));", "score": 0.9120715856552124 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/CCIndexFetcher.java", "retrieved_chunk": " new ObjectMapper().readValue(new File(args[0]), ExtractorConfig.class);\n execute(fetcherConfig);\n }\n private static void execute(ExtractorConfig fetcherConfig) throws Exception {\n ArrayBlockingQueue<FetchEmitTuple> indexPathsList = new ArrayBlockingQueue<>(1000);\n //IndexPathsReader reads a file containing a list of cc-index.paths files\n //and writes the literal gz files (cc-index/collections/CC-MAIN-2023-06/indexes/cdx-00000.gz)\n //to indexPathsList\n int totalThreads = fetcherConfig.getNumThreads() + 1;\n ExecutorService executorService = Executors.newFixedThreadPool(totalThreads);", "score": 0.8759914636611938 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/CCMimeCounter.java", "retrieved_chunk": " while (finishedWorkers < totalThreads) {\n //blocking\n Future<Long> future = executorCompletionService.take();\n if (future != null) {\n Long f = future.get();\n LOGGER.debug(\"completed {}\", f);\n if (f.equals(INDEX_WORKER_ID)) {\n finishedWorkers++;\n } else if (f.equals(INDEX_READER_ID)) {\n LOGGER.info(\"Index paths reader successfully completed\");", "score": 0.8712623119354248 } ]
java
finishedWorkers < fetcherConfig.getNumThreads()) {
/* * Licensed to the Apache Software Foundation (ASF) under one or more * contributor license agreements. See the NOTICE file distributed with * this work for additional information regarding copyright ownership. * The ASF licenses this file to You under the Apache License, Version 2.0 * (the "License"); you may not use this file except in compliance with * the License. You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ package org.tallison.cc.index.extractor; import java.io.File; import java.io.IOException; import java.io.InputStream; import java.util.Collections; import java.util.concurrent.ArrayBlockingQueue; import java.util.concurrent.Callable; import java.util.concurrent.ExecutionException; import java.util.concurrent.ExecutorCompletionService; import java.util.concurrent.ExecutorService; import java.util.concurrent.Executors; import java.util.concurrent.Future; import java.util.concurrent.TimeUnit; import java.util.concurrent.TimeoutException; import com.fasterxml.jackson.databind.ObjectMapper; import org.slf4j.Logger; import org.slf4j.LoggerFactory; import org.tallison.cc.index.IndexIterator; import org.apache.tika.exception.TikaException; import org.apache.tika.metadata.Metadata; import org.apache.tika.pipes.FetchEmitTuple; import org.apache.tika.pipes.emitter.StreamEmitter; import org.apache.tika.pipes.fetcher.Fetcher; import org.apache.tika.pipes.pipesiterator.CallablePipesIterator; import org.apache.tika.pipes.pipesiterator.PipesIterator; /** * This class fetches index files from aws to a local file share. * <p> * This pulls the index files either via https or s3 */ public class CCIndexFetcher { private static final Logger LOGGER = LoggerFactory.getLogger(CCIndexFetcher.class); public static void main(String[] args) throws Exception { ExtractorConfig fetcherConfig = new ObjectMapper().readValue(new File(args[0]), ExtractorConfig.class); execute(fetcherConfig); } private static void execute(ExtractorConfig fetcherConfig) throws Exception { ArrayBlockingQueue<FetchEmitTuple> indexPathsList = new ArrayBlockingQueue<>(1000); //IndexPathsReader reads a file containing a list of cc-index.paths files //and writes the literal gz files (cc-index/collections/CC-MAIN-2023-06/indexes/cdx-00000.gz) //to indexPathsList int totalThreads = fetcherConfig.getNumThreads() + 1; ExecutorService executorService = Executors.newFixedThreadPool(totalThreads); ExecutorCompletionService<Long> executorCompletionService = new ExecutorCompletionService<>(executorService); IndexIterator indexIterator = fetcherConfig.getIndexIterator(); indexIterator.initialize(Collections.EMPTY_MAP); executorCompletionService.submit(new CallablePipesIterator(indexIterator, indexPathsList)); int finishedWorkers = 0; try { for (int i = 0; i < fetcherConfig.getNumThreads(); i++) { executorCompletionService.submit(new IndexFetcher(fetcherConfig, indexPathsList)); } while (finishedWorkers < totalThreads) { //blocking Future<Long> future = executorCompletionService.take(); if (future != null) { Long f = future.get(); finishedWorkers++; LOGGER.debug("completed {}: {}", f, finishedWorkers); } } } catch (ExecutionException e) { LOGGER.error("main loop exception", e); throw new RuntimeException(e); } catch (InterruptedException e) { LOGGER.warn("main loop interrupted exception", e); throw new RuntimeException(e); } finally { executorService.shutdown(); executorService.shutdownNow(); } } private static class IndexFetcher implements Callable<Long> { private final ExtractorConfig fetcherConfig; private final ArrayBlockingQueue<FetchEmitTuple> indexPathsList; public IndexFetcher(ExtractorConfig fetcherConfig, ArrayBlockingQueue<FetchEmitTuple> indexPathsList) { this.fetcherConfig = fetcherConfig; this.indexPathsList = indexPathsList; } @Override public Long call() throws Exception { Fetcher fetcher = fetcherConfig.newFetcher(); StreamEmitter
streamEmitter = fetcherConfig.newEmitter();
while (true) { FetchEmitTuple t = indexPathsList.poll(120, TimeUnit.MINUTES); if (t == null) { throw new TimeoutException("waited 120 minutes for a new record"); } if (t == PipesIterator.COMPLETED_SEMAPHORE) { indexPathsList.put(PipesIterator.COMPLETED_SEMAPHORE); LOGGER.info("Index fetcher finished"); return 1l; } fetch(t, fetcher, streamEmitter); } } private void fetch(FetchEmitTuple t, Fetcher fetcher, StreamEmitter streamEmitter) { LOGGER.info("about to download: " + t.getFetchKey().getFetchKey()); try (InputStream is = fetcher.fetch(t.getFetchKey().getFetchKey(), new Metadata())) { streamEmitter.emit(t.getFetchKey().getFetchKey(), is, new Metadata()); LOGGER.info("successfully downloaded: " + t.getFetchKey().getFetchKey()); } catch (TikaException | IOException e) { LOGGER.error("failed to copy " + t.getFetchKey().getFetchKey(), e); } } } }
src/main/java/org/tallison/cc/index/extractor/CCIndexFetcher.java
tballison-commoncrawl-fetcher-lite-2c106d1
[ { "filename": "src/main/java/org/tallison/cc/index/extractor/CCMimeCounter.java", "retrieved_chunk": " private final Fetcher fetcher;\n IndexWorker(ExtractorConfig fetcherConfig, ArrayBlockingQueue<FetchEmitTuple> indexUrls,\n AbstractRecordProcessor recordProcessor) throws TikaException {\n this.indexUrls = indexUrls;\n this.recordProcessor = recordProcessor;\n this.fetcher = fetcherConfig.newFetcher();\n }\n @Override\n public Long call() throws Exception {\n boolean shouldContinue = true;", "score": 0.9175136089324951 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/CCFileExtractor.java", "retrieved_chunk": " this.indexUrls = indexUrls;\n this.recordProcessor = recordProcessor;\n this.indexFetcher = fetcherConfig.newIndexFetcher();\n }\n @Override\n public Long call() throws Exception {\n boolean shouldContinue = true;\n while (shouldContinue) {\n FetchEmitTuple indexUrl = indexUrls.poll(120, TimeUnit.MINUTES);\n if (indexUrl == null) {", "score": 0.9051171541213989 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/CCFileExtractor.java", "retrieved_chunk": " executorService.shutdown();\n executorService.shutdownNow();\n }\n }\n private static class IndexWorker implements Callable<Long> {\n private final ArrayBlockingQueue<FetchEmitTuple> indexUrls;\n private final AbstractRecordProcessor recordProcessor;\n private final Fetcher indexFetcher;\n IndexWorker(ExtractorConfig fetcherConfig, ArrayBlockingQueue<FetchEmitTuple> indexUrls,\n AbstractRecordProcessor recordProcessor) throws TikaException {", "score": 0.8958463072776794 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/ExtractorConfig.java", "retrieved_chunk": " }\n return fetchConfig.newFetcher();\n }\n public Fetcher newIndexFetcher() throws TikaConfigException {\n if (indexFetchConfig == null) {\n indexFetchConfig = new FetchConfig(null, DEFAULT_THROTTLE_SECONDS, null);\n }\n return indexFetchConfig.newFetcher();\n }\n public StreamEmitter newEmitter() throws TikaConfigException {", "score": 0.876887321472168 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/CCFileExtractor.java", "retrieved_chunk": " ExtractorConfig fetcherConfig =\n new ObjectMapper().readValue(new File(args[0]), ExtractorConfig.class);\n execute(fetcherConfig);\n }\n private static void execute(ExtractorConfig fetcherConfig) throws TikaException {\n ArrayBlockingQueue<FetchEmitTuple> indexPathsList = new ArrayBlockingQueue<>(1000);\n //IndexPathsReader reads a file containing a list of cc-index.paths files\n //and writes the literal gz files (cc-index/collections/CC-MAIN-2023-06/indexes/cdx-00000.gz)\n //to indexPathsList\n //IndexWorker reads a single index.gz file at a time and processes each record", "score": 0.8752036094665527 } ]
java
streamEmitter = fetcherConfig.newEmitter();
package com.tcoded.folialib.impl; import com.tcoded.folialib.FoliaLib; import com.tcoded.folialib.enums.EntityTaskResult; import com.tcoded.folialib.util.TimeConverter; import com.tcoded.folialib.wrapper.WrappedTask; import com.tcoded.folialib.wrapper.task.WrappedFoliaTask; import io.papermc.paper.threadedregions.scheduler.AsyncScheduler; import io.papermc.paper.threadedregions.scheduler.GlobalRegionScheduler; import org.bukkit.Chunk; import org.bukkit.Location; import org.bukkit.World; import org.bukkit.entity.Entity; import org.bukkit.entity.Player; import org.bukkit.plugin.java.JavaPlugin; import java.util.UUID; import java.util.concurrent.CompletableFuture; import java.util.concurrent.TimeUnit; public class FoliaImplementation implements ServerImplementation { private final JavaPlugin plugin; private final GlobalRegionScheduler globalRegionScheduler; private final AsyncScheduler asyncScheduler; public FoliaImplementation(FoliaLib foliaLib) { this.plugin = foliaLib.getPlugin(); this.globalRegionScheduler = plugin.getServer().getGlobalRegionScheduler(); this.asyncScheduler = plugin.getServer().getAsyncScheduler(); } @Override public CompletableFuture<Void> runNextTick(Runnable runnable) { CompletableFuture<Void> future = new CompletableFuture<>(); this.globalRegionScheduler.execute(plugin, () -> { runnable.run(); future.complete(null); }); return future; } @Override public CompletableFuture<Void> runAsync(Runnable runnable) { CompletableFuture<Void> future = new CompletableFuture<>(); this.asyncScheduler.runNow(plugin, task -> { runnable.run(); future.complete(null); }); return future; } @Override public WrappedTask runLater(Runnable runnable, long delay, TimeUnit unit) { return new WrappedFoliaTask( this.globalRegionScheduler.runDelayed( plugin, task -> runnable.
run(), TimeConverter.toTicks(delay, unit) ) );
} @Override public WrappedTask runLaterAsync(Runnable runnable, long delay, TimeUnit unit) { return new WrappedFoliaTask( this.asyncScheduler.runDelayed( plugin, task -> runnable.run(), delay, unit ) ); } @Override public WrappedTask runTimer(Runnable runnable, long delay, long period, TimeUnit unit) { return new WrappedFoliaTask( this.globalRegionScheduler.runAtFixedRate( plugin, task -> runnable.run(), TimeConverter.toTicks(delay, unit), TimeConverter.toTicks(period, unit) ) ); } @Override public WrappedTask runTimerAsync(Runnable runnable, long delay, long period, TimeUnit unit) { return new WrappedFoliaTask( this.asyncScheduler.runAtFixedRate( plugin, task -> runnable.run(), delay, period, unit ) ); } @Override public CompletableFuture<Void> runAtLocation(Location location, Runnable runnable) { CompletableFuture<Void> future = new CompletableFuture<>(); this.plugin.getServer().getRegionScheduler().execute(plugin, location, () -> { runnable.run(); future.complete(null); }); return future; } @Override public WrappedTask runAtLocationLater(Location location, Runnable runnable, long delay, TimeUnit unit) { return new WrappedFoliaTask( this.plugin.getServer().getRegionScheduler().runDelayed( plugin, location, task -> runnable.run(), TimeConverter.toTicks(delay, unit) ) ); } @Override public WrappedTask runAtLocationTimer(Location location, Runnable runnable, long delay, long period, TimeUnit unit) { return new WrappedFoliaTask( this.plugin.getServer().getRegionScheduler().runAtFixedRate( plugin, location, task -> runnable.run(), TimeConverter.toTicks(delay, unit), TimeConverter.toTicks(period, unit) ) ); } @Override public CompletableFuture<EntityTaskResult> runAtEntity(Entity entity, Runnable runnable) { CompletableFuture<EntityTaskResult> future = new CompletableFuture<>(); boolean success = entity.getScheduler().execute(this.plugin, () -> { runnable.run(); future.complete(EntityTaskResult.SUCCESS); }, null, 0); if (!success) { future.complete(EntityTaskResult.SCHEDULER_RETIRED); } return future; } @Override public CompletableFuture<EntityTaskResult> runAtEntityWithFallback(Entity entity, Runnable runnable, Runnable fallback) { CompletableFuture<EntityTaskResult> future = new CompletableFuture<>(); boolean success = entity.getScheduler().execute(this.plugin, () -> { runnable.run(); future.complete(EntityTaskResult.SUCCESS); }, () -> { fallback.run(); future.complete(EntityTaskResult.ENTITY_RETIRED); }, 0); if (!success) { future.complete(EntityTaskResult.SCHEDULER_RETIRED); } return future; } @Override public WrappedTask runAtEntityLater(Entity entity, Runnable runnable, long delay, TimeUnit unit) { return new WrappedFoliaTask( entity.getScheduler().runDelayed( plugin, task -> runnable.run(), null, TimeConverter.toTicks(delay, unit) ) ); } @Override public WrappedTask runAtEntityTimer(Entity entity, Runnable runnable, long delay, long period, TimeUnit unit) { return new WrappedFoliaTask( entity.getScheduler().runAtFixedRate( plugin, task -> runnable.run(), null, TimeConverter.toTicks(delay, unit), TimeConverter.toTicks(period, unit) ) ); } @Override public void cancelTask(WrappedTask task) { task.cancel(); } @Override public void cancelAllTasks() { this.globalRegionScheduler.cancelTasks(plugin); this.asyncScheduler.cancelTasks(plugin); } @Override public Player getPlayer(String name) { // This is thread-safe in folia return this.plugin.getServer().getPlayer(name); } @Override public Player getPlayerExact(String name) { // This is thread-safe in folia return this.plugin.getServer().getPlayerExact(name); } @Override public Player getPlayer(UUID uuid) { // This is thread-safe in folia return this.plugin.getServer().getPlayer(uuid); } @Override public CompletableFuture<Boolean> teleportAsync(Player player, Location location) { return player.teleportAsync(location); } }
src/main/java/com/tcoded/folialib/impl/FoliaImplementation.java
TechnicallyCoded-FoliaLib-8f9f24f
[ { "filename": "src/main/java/com/tcoded/folialib/impl/SpigotImplementation.java", "retrieved_chunk": " return future;\n }\n @Override\n public WrappedTask runLater(Runnable runnable, long delay, TimeUnit unit) {\n return new WrappedBukkitTask(\n this.scheduler.runTaskLater(plugin, runnable, TimeConverter.toTicks(delay, unit))\n );\n }\n @Override\n public WrappedTask runLaterAsync(Runnable runnable, long delay, TimeUnit unit) {", "score": 0.9420515298843384 }, { "filename": "src/main/java/com/tcoded/folialib/impl/ServerImplementation.java", "retrieved_chunk": " * Folia: Synced with the server daylight cycle tick\n * Paper: Synced with the server main thread\n * Spigot: Synced with the server main thread\n * @param runnable Task to run\n * @param delay Delay before execution\n * @param unit Time unit\n * @return WrappedTask instance\n */\n WrappedTask runLater(Runnable runnable, long delay, TimeUnit unit);\n /**", "score": 0.8927702903747559 }, { "filename": "src/main/java/com/tcoded/folialib/impl/ServerImplementation.java", "retrieved_chunk": " * Folia: Async\n * Paper: Async\n * Spigot: Async\n * @param runnable Task to run\n * @param delay Delay before execution\n * @param unit Time unit\n * @return WrappedTask instance\n */\n WrappedTask runLaterAsync(Runnable runnable, long delay, TimeUnit unit);\n // ----- Global Timers -----", "score": 0.8923441171646118 }, { "filename": "src/main/java/com/tcoded/folialib/impl/SpigotImplementation.java", "retrieved_chunk": " return new WrappedBukkitTask(\n this.scheduler.runTaskLaterAsynchronously(plugin, runnable, TimeConverter.toTicks(delay, unit))\n );\n }\n @Override\n public WrappedTask runTimer(Runnable runnable, long delay, long period, TimeUnit unit) {\n return new WrappedBukkitTask(\n this.scheduler.runTaskTimer(\n plugin, runnable,\n TimeConverter.toTicks(delay, unit),", "score": 0.8867412805557251 }, { "filename": "src/main/java/com/tcoded/folialib/impl/SpigotImplementation.java", "retrieved_chunk": " }\n @Override\n public WrappedTask runAtLocationLater(Location location, Runnable runnable, long delay, TimeUnit unit) {\n return new WrappedBukkitTask(\n this.scheduler.runTaskLater(plugin, runnable, TimeConverter.toTicks(delay, unit))\n );\n }\n @Override\n public WrappedTask runAtLocationTimer(Location location, Runnable runnable, long delay, long period, TimeUnit unit) {\n return new WrappedBukkitTask(", "score": 0.8791068196296692 } ]
java
run(), TimeConverter.toTicks(delay, unit) ) );
/* * Licensed to the Apache Software Foundation (ASF) under one or more * contributor license agreements. See the NOTICE file distributed with * this work for additional information regarding copyright ownership. * The ASF licenses this file to You under the Apache License, Version 2.0 * (the "License"); you may not use this file except in compliance with * the License. You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ package org.tallison.cc.index.extractor; import java.io.ByteArrayInputStream; import java.io.ByteArrayOutputStream; import java.io.IOException; import java.io.InputStream; import java.nio.file.Files; import java.nio.file.Path; import java.nio.file.StandardCopyOption; import java.util.Optional; import java.util.zip.GZIPInputStream; import org.apache.commons.codec.binary.Base32; import org.apache.commons.codec.digest.DigestUtils; import org.apache.commons.io.IOUtils; import org.netpreserve.jwarc.MediaType; import org.netpreserve.jwarc.WarcPayload; import org.netpreserve.jwarc.WarcReader; import org.netpreserve.jwarc.WarcRecord; import org.netpreserve.jwarc.WarcResponse; import org.slf4j.Logger; import org.slf4j.LoggerFactory; import org.tallison.cc.index.CCIndexReaderCounter; import org.tallison.cc.index.CCIndexRecord; import org.tallison.cc.index.io.TargetPathRewriter; import org.apache.tika.exception.TikaConfigException; import org.apache.tika.exception.TikaException; import org.apache.tika.io.TikaInputStream; import org.apache.tika.metadata.Metadata; import org.apache.tika.pipes.FetchEmitTuple; import org.apache.tika.pipes.emitter.EmitKey; import org.apache.tika.pipes.emitter.StreamEmitter; import org.apache.tika.pipes.fetcher.FetchKey; import org.apache.tika.pipes.fetcher.RangeFetcher; public class FileFromCCWarcExtractor { private static Logger LOGGER = LoggerFactory.getLogger(FileFromCCWarcExtractor.class); private static Logger EXTRACTED_LOGGER = LoggerFactory.getLogger("extracted-urls"); private static Logger EXTRACTED_ALL_LOGGER = LoggerFactory.getLogger("extracted-urls-all"); private final StreamEmitter emitter; private final TargetPathRewriter targetPathRewriter; private RangeFetcher fetcher; private final boolean extractTruncated; private Base32 base32 = new Base32(); private final CCIndexReaderCounter ccIndexReaderCounter; public FileFromCCWarcExtractor(ExtractorConfig fetcherConfig, CCIndexReaderCounter ccIndexReaderCounter) throws TikaConfigException { this.emitter = fetcherConfig.newEmitter(); this.fetcher = (RangeFetcher) fetcherConfig.newFetcher(); this.targetPathRewriter = fetcherConfig.getTargetPathRewriter(); this.extractTruncated = fetcherConfig.isExtractTruncated(); this.ccIndexReaderCounter = ccIndexReaderCounter; } public void fetchToPath(CCIndexRecord record) throws InterruptedException { LOGGER.debug("going to fetch {} {}->{}"
, record.getFilename(), record.getOffset(), record.getLength());
FetchEmitTuple t = new FetchEmitTuple(record.getFilename(), new FetchKey("", record.getFilename(), record.getOffset(), record.getOffset() + record.getLength() - 1), new EmitKey()); byte[] warcRecordGZBytes; try { warcRecordGZBytes = fetchWarcBytes(t); } catch (TikaException | IOException e) { LOGGER.warn("couldn't get bytes from cc's warc " + t, e); return; } String id = record.getUrl(); try { parseWarc(id, record, warcRecordGZBytes); } catch (IOException e) { LOGGER.warn("problem parsing warc file", e); } } private void fetchPayload(String id, CCIndexRecord ccIndexRecord, WarcRecord record) throws IOException { if (!((record instanceof WarcResponse) && record.contentType().base().equals(MediaType.HTTP))) { return; } Optional<WarcPayload> payload = ((WarcResponse) record).payload(); if (!payload.isPresent()) { LOGGER.debug("no payload {}", id); ccIndexReaderCounter.getEmptyPayload().incrementAndGet(); return; } if (payload.get().body().size() == 0) { LOGGER.debug("empty payload id={}", id); ccIndexReaderCounter.getEmptyPayload().incrementAndGet(); return; } Path tmp = Files.createTempFile("ccfile-fetcher-", ""); try { Files.copy(payload.get().body().stream(), tmp, StandardCopyOption.REPLACE_EXISTING); String targetDigest = null; String base32Sha1 = ""; try (InputStream is = Files.newInputStream(tmp)) { base32Sha1 = base32.encodeAsString(DigestUtils.sha1(is)); } catch (IOException e) { LOGGER.warn("IOException during digesting: " + tmp.toAbsolutePath()); return; } if (!base32Sha1.equals(ccIndexRecord.getDigest())) { LOGGER.warn("Bad digest for url={} ccindex={} sha1={}", id, ccIndexRecord.getDigest(), base32Sha1); } //TODO: make digest and encoding configurable try (InputStream is = Files.newInputStream(tmp)) { targetDigest = DigestUtils.sha256Hex(is); } catch (IOException e) { LOGGER.warn("IOException during digesting: " + tmp.toAbsolutePath()); return; } long length = -1; try { length = Files.size(tmp); } catch (IOException e) { LOGGER.warn("IOException during digesting: " + tmp.toAbsolutePath()); return; } String targetPath = targetPathRewriter.rewrite(targetDigest); Metadata metadata = new Metadata(); try (InputStream is = TikaInputStream.get(tmp, metadata)) { emitter.emit(targetPath, is, new Metadata()); logSuccess(ccIndexRecord, targetDigest, length, targetPath); } catch (IOException | TikaException e) { LOGGER.warn("problem writing id={}", id, e); } } finally { try { Files.delete(tmp); } catch (IOException e) { LOGGER.warn("can't delete " + tmp.toAbsolutePath(), e); } } } private void logSuccess(CCIndexRecord ccIndexRecord, String targetDigest, long length, String targetPath) { if (extractTruncated) { EXTRACTED_ALL_LOGGER.info("", ccIndexRecord.getUrl(), ccIndexRecord.getNormalizedMime(), ccIndexRecord.getNormalizedMimeDetected(), ccIndexRecord.getFilename(), ccIndexRecord.getOffset(), ccIndexRecord.getLength(), ccIndexRecord.getTruncated(), targetDigest, length, targetPath); } else { //new ObjectArray ? //url,mime_detected,warc_file,warc_offset,warc_length,sha256,length,path EXTRACTED_LOGGER.info("", ccIndexRecord.getUrl(), ccIndexRecord.getNormalizedMime(), ccIndexRecord.getNormalizedMimeDetected(), ccIndexRecord.getFilename(), ccIndexRecord.getOffset(), ccIndexRecord.getLength(), targetDigest, length, targetPath); } } private void parseWarc(String id, CCIndexRecord ccIndexRecord, byte[] warcRecordGZBytes) throws IOException { //need to leave initial inputstream open while parsing warcrecord //can't just parse record and return try (InputStream is = new GZIPInputStream(new ByteArrayInputStream(warcRecordGZBytes))) { try (WarcReader warcreader = new WarcReader(is)) { //should be a single warc per file //return the first for (WarcRecord warcRecord : warcreader) { fetchPayload(id, ccIndexRecord, warcRecord); return; } } } } private byte[] fetchWarcBytes(FetchEmitTuple t) throws TikaException, InterruptedException, IOException { ByteArrayOutputStream bos = new ByteArrayOutputStream(); FetchKey k = t.getFetchKey(); try (InputStream is = fetcher.fetch(k.getFetchKey(), k.getRangeStart(), k.getRangeEnd(), new Metadata())) { IOUtils.copy(is, bos); } return bos.toByteArray(); } }
src/main/java/org/tallison/cc/index/extractor/FileFromCCWarcExtractor.java
tballison-commoncrawl-fetcher-lite-2c106d1
[ { "filename": "src/main/java/org/tallison/cc/index/extractor/CCMimeCounter.java", "retrieved_chunk": " private final Fetcher fetcher;\n IndexWorker(ExtractorConfig fetcherConfig, ArrayBlockingQueue<FetchEmitTuple> indexUrls,\n AbstractRecordProcessor recordProcessor) throws TikaException {\n this.indexUrls = indexUrls;\n this.recordProcessor = recordProcessor;\n this.fetcher = fetcherConfig.newFetcher();\n }\n @Override\n public Long call() throws Exception {\n boolean shouldContinue = true;", "score": 0.8485296964645386 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/CCFileExtractor.java", "retrieved_chunk": " this.indexUrls = indexUrls;\n this.recordProcessor = recordProcessor;\n this.indexFetcher = fetcherConfig.newIndexFetcher();\n }\n @Override\n public Long call() throws Exception {\n boolean shouldContinue = true;\n while (shouldContinue) {\n FetchEmitTuple indexUrl = indexUrls.poll(120, TimeUnit.MINUTES);\n if (indexUrl == null) {", "score": 0.848323404788971 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/CCFileExtractor.java", "retrieved_chunk": " }\n return INDEX_WORKER_ID;\n }\n private boolean processFile(FetchEmitTuple fetchEmitTuple,\n AbstractRecordProcessor recordProcessor)\n throws InterruptedException {\n long start = System.currentTimeMillis();\n LOGGER.info(\"starting to fetch index gz: {}\",\n fetchEmitTuple.getFetchKey().getFetchKey());\n try (TikaInputStream tis = (TikaInputStream) indexFetcher.fetch(", "score": 0.8425893783569336 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/CCFileExtractorRecordProcessor.java", "retrieved_chunk": " LoggerFactory.getLogger(\"truncated-urls-full\");\n private final ExtractorConfig fetcherConfig;\n private final CCIndexReaderCounter counter;\n private final FileFromCCWarcExtractor fileFromCCWarcFetcher;\n private long reportEvery = 100000;\n public CCFileExtractorRecordProcessor(ExtractorConfig fetcherConfig, CCIndexReaderCounter counter)\n throws TikaConfigException, IOException {\n this.fetcherConfig = fetcherConfig;\n this.counter = counter;\n this.fileFromCCWarcFetcher = new FileFromCCWarcExtractor(fetcherConfig, counter);", "score": 0.8401328325271606 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/CCIndexFetcher.java", "retrieved_chunk": " new ObjectMapper().readValue(new File(args[0]), ExtractorConfig.class);\n execute(fetcherConfig);\n }\n private static void execute(ExtractorConfig fetcherConfig) throws Exception {\n ArrayBlockingQueue<FetchEmitTuple> indexPathsList = new ArrayBlockingQueue<>(1000);\n //IndexPathsReader reads a file containing a list of cc-index.paths files\n //and writes the literal gz files (cc-index/collections/CC-MAIN-2023-06/indexes/cdx-00000.gz)\n //to indexPathsList\n int totalThreads = fetcherConfig.getNumThreads() + 1;\n ExecutorService executorService = Executors.newFixedThreadPool(totalThreads);", "score": 0.8382256627082825 } ]
java
, record.getFilename(), record.getOffset(), record.getLength());
package com.tcoded.folialib.impl; import com.tcoded.folialib.FoliaLib; import com.tcoded.folialib.enums.EntityTaskResult; import com.tcoded.folialib.util.TimeConverter; import com.tcoded.folialib.wrapper.WrappedTask; import com.tcoded.folialib.wrapper.task.WrappedBukkitTask; import org.bukkit.Location; import org.bukkit.entity.Entity; import org.bukkit.entity.Player; import org.bukkit.plugin.java.JavaPlugin; import org.bukkit.scheduler.BukkitScheduler; import org.jetbrains.annotations.NotNull; import java.util.UUID; import java.util.concurrent.CompletableFuture; import java.util.concurrent.ExecutionException; import java.util.concurrent.TimeUnit; public class SpigotImplementation implements ServerImplementation { private final JavaPlugin plugin; @SuppressWarnings("deprecation") private final @NotNull BukkitScheduler scheduler; public SpigotImplementation(FoliaLib foliaLib) { this.plugin = foliaLib.getPlugin(); this.scheduler = plugin.getServer().getScheduler(); } @Override public CompletableFuture<Void> runNextTick(Runnable runnable) { CompletableFuture<Void> future = new CompletableFuture<>(); this.scheduler.runTask(plugin, () -> { runnable.run(); future.complete(null); }); return future; } @Override public CompletableFuture<Void> runAsync(Runnable runnable) { CompletableFuture<Void> future = new CompletableFuture<>(); this.scheduler.runTaskAsynchronously(plugin, () -> { runnable.run(); future.complete(null); }); return future; } @Override public WrappedTask runLater(Runnable runnable, long delay, TimeUnit unit) { return new WrappedBukkitTask( this.scheduler.runTaskLater(plugin, runnable, TimeConverter.toTicks(delay, unit)) ); } @Override public WrappedTask runLaterAsync(Runnable runnable, long delay, TimeUnit unit) { return new WrappedBukkitTask( this.scheduler.runTaskLaterAsynchronously(plugin, runnable, TimeConverter.toTicks(delay, unit)) ); } @Override public WrappedTask runTimer(Runnable runnable, long delay, long period, TimeUnit unit) { return new WrappedBukkitTask( this.scheduler.runTaskTimer( plugin, runnable, TimeConverter.toTicks(delay, unit),
TimeConverter.toTicks(period, unit)) );
} @Override public WrappedTask runTimerAsync(Runnable runnable, long delay, long period, TimeUnit unit) { return new WrappedBukkitTask( this.scheduler.runTaskTimerAsynchronously( plugin, runnable, TimeConverter.toTicks(delay, unit), TimeConverter.toTicks(period, unit)) ); } @Override public CompletableFuture<Void> runAtLocation(Location location, Runnable runnable) { CompletableFuture<Void> future = new CompletableFuture<>(); this.scheduler.runTask(plugin, () -> { runnable.run(); future.complete(null); }); return future; } @Override public WrappedTask runAtLocationLater(Location location, Runnable runnable, long delay, TimeUnit unit) { return new WrappedBukkitTask( this.scheduler.runTaskLater(plugin, runnable, TimeConverter.toTicks(delay, unit)) ); } @Override public WrappedTask runAtLocationTimer(Location location, Runnable runnable, long delay, long period, TimeUnit unit) { return new WrappedBukkitTask( this.scheduler.runTaskTimer( plugin, runnable, TimeConverter.toTicks(delay, unit), TimeConverter.toTicks(period, unit)) ); } @Override public CompletableFuture<EntityTaskResult> runAtEntity(Entity entity, Runnable runnable) { CompletableFuture<EntityTaskResult> future = new CompletableFuture<>(); this.scheduler.runTask(plugin, () -> { runnable.run(); future.complete(EntityTaskResult.SUCCESS); }); return future; } @Override public CompletableFuture<EntityTaskResult> runAtEntityWithFallback(Entity entity, Runnable runnable, Runnable fallback) { CompletableFuture<EntityTaskResult> future = new CompletableFuture<>(); this.scheduler.runTask(plugin, () -> { if (entity.isValid()) { runnable.run(); future.complete(EntityTaskResult.SUCCESS); } else { fallback.run(); future.complete(EntityTaskResult.ENTITY_RETIRED); } }); return future; } @Override public WrappedTask runAtEntityLater(Entity entity, Runnable runnable, long delay, TimeUnit unit) { return new WrappedBukkitTask( this.scheduler.runTaskLater(plugin, runnable, TimeConverter.toTicks(delay, unit)) ); } @Override public WrappedTask runAtEntityTimer(Entity entity, Runnable runnable, long delay, long period, TimeUnit unit) { return new WrappedBukkitTask( this.scheduler.runTaskTimer( plugin, runnable, TimeConverter.toTicks(delay, unit), TimeConverter.toTicks(period, unit)) ); } @Override public void cancelTask(WrappedTask task) { task.cancel(); } @Override public void cancelAllTasks() { this.scheduler.cancelTasks(plugin); } @Override public Player getPlayer(String name) { // Already on the main thread if (this.plugin.getServer().isPrimaryThread()) { return this.plugin.getServer().getPlayer(name); } // Not on the main thread, we need to wait until the next tick else { try { return this.scheduler.callSyncMethod(plugin, () -> this.plugin.getServer().getPlayer(name)).get(); } catch (InterruptedException | ExecutionException e) { e.printStackTrace(); } } // Fallback to null return null; } @Override public Player getPlayerExact(String name) { // Already on the main thread if (this.plugin.getServer().isPrimaryThread()) { return this.plugin.getServer().getPlayerExact(name); } // Not on the main thread, we need to wait until the next tick else { try { return this.scheduler.callSyncMethod(plugin, () -> this.plugin.getServer().getPlayerExact(name)).get(); } catch (InterruptedException | ExecutionException e) { e.printStackTrace(); } } // Fallback to null return null; } @SuppressWarnings("DuplicatedCode") @Override public Player getPlayer(UUID uuid) { // Already on the main thread if (this.plugin.getServer().isPrimaryThread()) { return this.plugin.getServer().getPlayer(uuid); } // Not on the main thread, we need to wait until the next tick else { try { return this.scheduler.callSyncMethod(plugin, () -> this.plugin.getServer().getPlayer(uuid)).get(); } catch (InterruptedException | ExecutionException e) { e.printStackTrace(); } } // Fallback to null return null; } @Override public CompletableFuture<Boolean> teleportAsync(Player player, Location location) { CompletableFuture<Boolean> future = new CompletableFuture<>(); this.runAtEntity(player, () -> { if (player.isValid() && player.isOnline()) { player.teleport(location); future.complete(true); } else { future.complete(false); } }); return future; } }
src/main/java/com/tcoded/folialib/impl/SpigotImplementation.java
TechnicallyCoded-FoliaLib-8f9f24f
[ { "filename": "src/main/java/com/tcoded/folialib/impl/FoliaImplementation.java", "retrieved_chunk": " @Override\n public WrappedTask runTimer(Runnable runnable, long delay, long period, TimeUnit unit) {\n return new WrappedFoliaTask(\n this.globalRegionScheduler.runAtFixedRate(\n plugin, task -> runnable.run(),\n TimeConverter.toTicks(delay, unit),\n TimeConverter.toTicks(period, unit)\n )\n );\n }", "score": 0.9389793872833252 }, { "filename": "src/main/java/com/tcoded/folialib/impl/FoliaImplementation.java", "retrieved_chunk": " @Override\n public WrappedTask runTimerAsync(Runnable runnable, long delay, long period, TimeUnit unit) {\n return new WrappedFoliaTask(\n this.asyncScheduler.runAtFixedRate(\n plugin, task -> runnable.run(),\n delay, period, unit\n )\n );\n }\n @Override", "score": 0.938437283039093 }, { "filename": "src/main/java/com/tcoded/folialib/impl/FoliaImplementation.java", "retrieved_chunk": " return new WrappedFoliaTask(\n this.plugin.getServer().getRegionScheduler().runDelayed(\n plugin, location, task -> runnable.run(),\n TimeConverter.toTicks(delay, unit)\n )\n );\n }\n @Override\n public WrappedTask runAtLocationTimer(Location location, Runnable runnable, long delay, long period, TimeUnit unit) {\n return new WrappedFoliaTask(", "score": 0.9115691184997559 }, { "filename": "src/main/java/com/tcoded/folialib/impl/FoliaImplementation.java", "retrieved_chunk": " null,\n TimeConverter.toTicks(delay, unit)\n )\n );\n }\n @Override\n public WrappedTask runAtEntityTimer(Entity entity, Runnable runnable, long delay, long period, TimeUnit unit) {\n return new WrappedFoliaTask(\n entity.getScheduler().runAtFixedRate(\n plugin,", "score": 0.9101240038871765 }, { "filename": "src/main/java/com/tcoded/folialib/impl/ServerImplementation.java", "retrieved_chunk": " WrappedTask runTimer(Runnable runnable, long delay, long period, TimeUnit unit);\n /**\n * Folia: Async\n * Paper: Async\n * Spigot: Async\n * @param runnable Task to run\n * @param delay Delay before first execution\n * @param period Delay between executions\n * @param unit Time unit\n * @return WrappedTask instance", "score": 0.9083977937698364 } ]
java
TimeConverter.toTicks(period, unit)) );
package com.tcoded.folialib.impl; import com.tcoded.folialib.FoliaLib; import com.tcoded.folialib.enums.EntityTaskResult; import com.tcoded.folialib.util.TimeConverter; import com.tcoded.folialib.wrapper.WrappedTask; import com.tcoded.folialib.wrapper.task.WrappedFoliaTask; import io.papermc.paper.threadedregions.scheduler.AsyncScheduler; import io.papermc.paper.threadedregions.scheduler.GlobalRegionScheduler; import org.bukkit.Chunk; import org.bukkit.Location; import org.bukkit.World; import org.bukkit.entity.Entity; import org.bukkit.entity.Player; import org.bukkit.plugin.java.JavaPlugin; import java.util.UUID; import java.util.concurrent.CompletableFuture; import java.util.concurrent.TimeUnit; public class FoliaImplementation implements ServerImplementation { private final JavaPlugin plugin; private final GlobalRegionScheduler globalRegionScheduler; private final AsyncScheduler asyncScheduler; public FoliaImplementation(FoliaLib foliaLib) { this.plugin = foliaLib.getPlugin(); this.globalRegionScheduler = plugin.getServer().getGlobalRegionScheduler(); this.asyncScheduler = plugin.getServer().getAsyncScheduler(); } @Override public CompletableFuture<Void> runNextTick(Runnable runnable) { CompletableFuture<Void> future = new CompletableFuture<>(); this.globalRegionScheduler.execute(plugin, () -> { runnable.run(); future.complete(null); }); return future; } @Override public CompletableFuture<Void> runAsync(Runnable runnable) { CompletableFuture<Void> future = new CompletableFuture<>(); this.asyncScheduler.runNow(plugin, task -> { runnable.run(); future.complete(null); }); return future; } @Override public WrappedTask runLater(Runnable runnable, long delay, TimeUnit unit) { return new WrappedFoliaTask( this.globalRegionScheduler.runDelayed( plugin, task -> runnable.run(), TimeConverter.toTicks(delay, unit) ) ); } @Override public WrappedTask runLaterAsync(Runnable runnable, long delay, TimeUnit unit) { return new WrappedFoliaTask( this.asyncScheduler.runDelayed( plugin, task -> runnable.run(), delay, unit ) ); } @Override public WrappedTask runTimer(Runnable runnable, long delay, long period, TimeUnit unit) { return new WrappedFoliaTask( this.globalRegionScheduler.runAtFixedRate( plugin, task -> runnable.run(),
TimeConverter.toTicks(delay, unit), TimeConverter.toTicks(period, unit) ) );
} @Override public WrappedTask runTimerAsync(Runnable runnable, long delay, long period, TimeUnit unit) { return new WrappedFoliaTask( this.asyncScheduler.runAtFixedRate( plugin, task -> runnable.run(), delay, period, unit ) ); } @Override public CompletableFuture<Void> runAtLocation(Location location, Runnable runnable) { CompletableFuture<Void> future = new CompletableFuture<>(); this.plugin.getServer().getRegionScheduler().execute(plugin, location, () -> { runnable.run(); future.complete(null); }); return future; } @Override public WrappedTask runAtLocationLater(Location location, Runnable runnable, long delay, TimeUnit unit) { return new WrappedFoliaTask( this.plugin.getServer().getRegionScheduler().runDelayed( plugin, location, task -> runnable.run(), TimeConverter.toTicks(delay, unit) ) ); } @Override public WrappedTask runAtLocationTimer(Location location, Runnable runnable, long delay, long period, TimeUnit unit) { return new WrappedFoliaTask( this.plugin.getServer().getRegionScheduler().runAtFixedRate( plugin, location, task -> runnable.run(), TimeConverter.toTicks(delay, unit), TimeConverter.toTicks(period, unit) ) ); } @Override public CompletableFuture<EntityTaskResult> runAtEntity(Entity entity, Runnable runnable) { CompletableFuture<EntityTaskResult> future = new CompletableFuture<>(); boolean success = entity.getScheduler().execute(this.plugin, () -> { runnable.run(); future.complete(EntityTaskResult.SUCCESS); }, null, 0); if (!success) { future.complete(EntityTaskResult.SCHEDULER_RETIRED); } return future; } @Override public CompletableFuture<EntityTaskResult> runAtEntityWithFallback(Entity entity, Runnable runnable, Runnable fallback) { CompletableFuture<EntityTaskResult> future = new CompletableFuture<>(); boolean success = entity.getScheduler().execute(this.plugin, () -> { runnable.run(); future.complete(EntityTaskResult.SUCCESS); }, () -> { fallback.run(); future.complete(EntityTaskResult.ENTITY_RETIRED); }, 0); if (!success) { future.complete(EntityTaskResult.SCHEDULER_RETIRED); } return future; } @Override public WrappedTask runAtEntityLater(Entity entity, Runnable runnable, long delay, TimeUnit unit) { return new WrappedFoliaTask( entity.getScheduler().runDelayed( plugin, task -> runnable.run(), null, TimeConverter.toTicks(delay, unit) ) ); } @Override public WrappedTask runAtEntityTimer(Entity entity, Runnable runnable, long delay, long period, TimeUnit unit) { return new WrappedFoliaTask( entity.getScheduler().runAtFixedRate( plugin, task -> runnable.run(), null, TimeConverter.toTicks(delay, unit), TimeConverter.toTicks(period, unit) ) ); } @Override public void cancelTask(WrappedTask task) { task.cancel(); } @Override public void cancelAllTasks() { this.globalRegionScheduler.cancelTasks(plugin); this.asyncScheduler.cancelTasks(plugin); } @Override public Player getPlayer(String name) { // This is thread-safe in folia return this.plugin.getServer().getPlayer(name); } @Override public Player getPlayerExact(String name) { // This is thread-safe in folia return this.plugin.getServer().getPlayerExact(name); } @Override public Player getPlayer(UUID uuid) { // This is thread-safe in folia return this.plugin.getServer().getPlayer(uuid); } @Override public CompletableFuture<Boolean> teleportAsync(Player player, Location location) { return player.teleportAsync(location); } }
src/main/java/com/tcoded/folialib/impl/FoliaImplementation.java
TechnicallyCoded-FoliaLib-8f9f24f
[ { "filename": "src/main/java/com/tcoded/folialib/impl/SpigotImplementation.java", "retrieved_chunk": " TimeConverter.toTicks(period, unit))\n );\n }\n @Override\n public WrappedTask runTimerAsync(Runnable runnable, long delay, long period, TimeUnit unit) {\n return new WrappedBukkitTask(\n this.scheduler.runTaskTimerAsynchronously(\n plugin, runnable,\n TimeConverter.toTicks(delay, unit),\n TimeConverter.toTicks(period, unit))", "score": 0.9280020594596863 }, { "filename": "src/main/java/com/tcoded/folialib/impl/SpigotImplementation.java", "retrieved_chunk": " return new WrappedBukkitTask(\n this.scheduler.runTaskLaterAsynchronously(plugin, runnable, TimeConverter.toTicks(delay, unit))\n );\n }\n @Override\n public WrappedTask runTimer(Runnable runnable, long delay, long period, TimeUnit unit) {\n return new WrappedBukkitTask(\n this.scheduler.runTaskTimer(\n plugin, runnable,\n TimeConverter.toTicks(delay, unit),", "score": 0.9151657223701477 }, { "filename": "src/main/java/com/tcoded/folialib/impl/ServerImplementation.java", "retrieved_chunk": " WrappedTask runTimer(Runnable runnable, long delay, long period, TimeUnit unit);\n /**\n * Folia: Async\n * Paper: Async\n * Spigot: Async\n * @param runnable Task to run\n * @param delay Delay before first execution\n * @param period Delay between executions\n * @param unit Time unit\n * @return WrappedTask instance", "score": 0.9054098129272461 }, { "filename": "src/main/java/com/tcoded/folialib/impl/ServerImplementation.java", "retrieved_chunk": " * @param period Delay between executions\n * @param unit Time unit\n * @return WrappedTask instance\n */\n WrappedTask runAtLocationTimer(Location location, Runnable runnable, long delay, long period, TimeUnit unit);\n // ----- Entity based -----\n /**\n * Folia: Synced with the tick of the region of the entity (even if the entity moves)\n * Paper: Synced with the server main thread\n * Spigot: Synced with the server main thread", "score": 0.8981728553771973 }, { "filename": "src/main/java/com/tcoded/folialib/impl/SpigotImplementation.java", "retrieved_chunk": " }\n @Override\n public WrappedTask runAtLocationLater(Location location, Runnable runnable, long delay, TimeUnit unit) {\n return new WrappedBukkitTask(\n this.scheduler.runTaskLater(plugin, runnable, TimeConverter.toTicks(delay, unit))\n );\n }\n @Override\n public WrappedTask runAtLocationTimer(Location location, Runnable runnable, long delay, long period, TimeUnit unit) {\n return new WrappedBukkitTask(", "score": 0.8915514945983887 } ]
java
TimeConverter.toTicks(delay, unit), TimeConverter.toTicks(period, unit) ) );
package com.tcoded.folialib.impl; import com.tcoded.folialib.FoliaLib; import com.tcoded.folialib.enums.EntityTaskResult; import com.tcoded.folialib.util.TimeConverter; import com.tcoded.folialib.wrapper.WrappedTask; import com.tcoded.folialib.wrapper.task.WrappedBukkitTask; import org.bukkit.Location; import org.bukkit.entity.Entity; import org.bukkit.entity.Player; import org.bukkit.plugin.java.JavaPlugin; import org.bukkit.scheduler.BukkitScheduler; import org.jetbrains.annotations.NotNull; import java.util.UUID; import java.util.concurrent.CompletableFuture; import java.util.concurrent.ExecutionException; import java.util.concurrent.TimeUnit; public class SpigotImplementation implements ServerImplementation { private final JavaPlugin plugin; @SuppressWarnings("deprecation") private final @NotNull BukkitScheduler scheduler; public SpigotImplementation(FoliaLib foliaLib) { this.plugin = foliaLib.getPlugin(); this.scheduler = plugin.getServer().getScheduler(); } @Override public CompletableFuture<Void> runNextTick(Runnable runnable) { CompletableFuture<Void> future = new CompletableFuture<>(); this.scheduler.runTask(plugin, () -> { runnable.run(); future.complete(null); }); return future; } @Override public CompletableFuture<Void> runAsync(Runnable runnable) { CompletableFuture<Void> future = new CompletableFuture<>(); this.scheduler.runTaskAsynchronously(plugin, () -> { runnable.run(); future.complete(null); }); return future; } @Override public WrappedTask runLater(Runnable runnable, long delay, TimeUnit unit) { return new WrappedBukkitTask( this.scheduler.runTaskLater(plugin, runnable, TimeConverter.toTicks(delay, unit)) ); } @Override public WrappedTask runLaterAsync(Runnable runnable, long delay, TimeUnit unit) { return new WrappedBukkitTask( this.scheduler.runTaskLaterAsynchronously(plugin, runnable, TimeConverter.toTicks(delay, unit)) ); } @Override public WrappedTask runTimer(Runnable runnable, long delay, long period, TimeUnit unit) { return new WrappedBukkitTask( this.scheduler.runTaskTimer( plugin, runnable,
TimeConverter.toTicks(delay, unit), TimeConverter.toTicks(period, unit)) );
} @Override public WrappedTask runTimerAsync(Runnable runnable, long delay, long period, TimeUnit unit) { return new WrappedBukkitTask( this.scheduler.runTaskTimerAsynchronously( plugin, runnable, TimeConverter.toTicks(delay, unit), TimeConverter.toTicks(period, unit)) ); } @Override public CompletableFuture<Void> runAtLocation(Location location, Runnable runnable) { CompletableFuture<Void> future = new CompletableFuture<>(); this.scheduler.runTask(plugin, () -> { runnable.run(); future.complete(null); }); return future; } @Override public WrappedTask runAtLocationLater(Location location, Runnable runnable, long delay, TimeUnit unit) { return new WrappedBukkitTask( this.scheduler.runTaskLater(plugin, runnable, TimeConverter.toTicks(delay, unit)) ); } @Override public WrappedTask runAtLocationTimer(Location location, Runnable runnable, long delay, long period, TimeUnit unit) { return new WrappedBukkitTask( this.scheduler.runTaskTimer( plugin, runnable, TimeConverter.toTicks(delay, unit), TimeConverter.toTicks(period, unit)) ); } @Override public CompletableFuture<EntityTaskResult> runAtEntity(Entity entity, Runnable runnable) { CompletableFuture<EntityTaskResult> future = new CompletableFuture<>(); this.scheduler.runTask(plugin, () -> { runnable.run(); future.complete(EntityTaskResult.SUCCESS); }); return future; } @Override public CompletableFuture<EntityTaskResult> runAtEntityWithFallback(Entity entity, Runnable runnable, Runnable fallback) { CompletableFuture<EntityTaskResult> future = new CompletableFuture<>(); this.scheduler.runTask(plugin, () -> { if (entity.isValid()) { runnable.run(); future.complete(EntityTaskResult.SUCCESS); } else { fallback.run(); future.complete(EntityTaskResult.ENTITY_RETIRED); } }); return future; } @Override public WrappedTask runAtEntityLater(Entity entity, Runnable runnable, long delay, TimeUnit unit) { return new WrappedBukkitTask( this.scheduler.runTaskLater(plugin, runnable, TimeConverter.toTicks(delay, unit)) ); } @Override public WrappedTask runAtEntityTimer(Entity entity, Runnable runnable, long delay, long period, TimeUnit unit) { return new WrappedBukkitTask( this.scheduler.runTaskTimer( plugin, runnable, TimeConverter.toTicks(delay, unit), TimeConverter.toTicks(period, unit)) ); } @Override public void cancelTask(WrappedTask task) { task.cancel(); } @Override public void cancelAllTasks() { this.scheduler.cancelTasks(plugin); } @Override public Player getPlayer(String name) { // Already on the main thread if (this.plugin.getServer().isPrimaryThread()) { return this.plugin.getServer().getPlayer(name); } // Not on the main thread, we need to wait until the next tick else { try { return this.scheduler.callSyncMethod(plugin, () -> this.plugin.getServer().getPlayer(name)).get(); } catch (InterruptedException | ExecutionException e) { e.printStackTrace(); } } // Fallback to null return null; } @Override public Player getPlayerExact(String name) { // Already on the main thread if (this.plugin.getServer().isPrimaryThread()) { return this.plugin.getServer().getPlayerExact(name); } // Not on the main thread, we need to wait until the next tick else { try { return this.scheduler.callSyncMethod(plugin, () -> this.plugin.getServer().getPlayerExact(name)).get(); } catch (InterruptedException | ExecutionException e) { e.printStackTrace(); } } // Fallback to null return null; } @SuppressWarnings("DuplicatedCode") @Override public Player getPlayer(UUID uuid) { // Already on the main thread if (this.plugin.getServer().isPrimaryThread()) { return this.plugin.getServer().getPlayer(uuid); } // Not on the main thread, we need to wait until the next tick else { try { return this.scheduler.callSyncMethod(plugin, () -> this.plugin.getServer().getPlayer(uuid)).get(); } catch (InterruptedException | ExecutionException e) { e.printStackTrace(); } } // Fallback to null return null; } @Override public CompletableFuture<Boolean> teleportAsync(Player player, Location location) { CompletableFuture<Boolean> future = new CompletableFuture<>(); this.runAtEntity(player, () -> { if (player.isValid() && player.isOnline()) { player.teleport(location); future.complete(true); } else { future.complete(false); } }); return future; } }
src/main/java/com/tcoded/folialib/impl/SpigotImplementation.java
TechnicallyCoded-FoliaLib-8f9f24f
[ { "filename": "src/main/java/com/tcoded/folialib/impl/FoliaImplementation.java", "retrieved_chunk": " @Override\n public WrappedTask runTimer(Runnable runnable, long delay, long period, TimeUnit unit) {\n return new WrappedFoliaTask(\n this.globalRegionScheduler.runAtFixedRate(\n plugin, task -> runnable.run(),\n TimeConverter.toTicks(delay, unit),\n TimeConverter.toTicks(period, unit)\n )\n );\n }", "score": 0.9399531483650208 }, { "filename": "src/main/java/com/tcoded/folialib/impl/FoliaImplementation.java", "retrieved_chunk": " @Override\n public WrappedTask runTimerAsync(Runnable runnable, long delay, long period, TimeUnit unit) {\n return new WrappedFoliaTask(\n this.asyncScheduler.runAtFixedRate(\n plugin, task -> runnable.run(),\n delay, period, unit\n )\n );\n }\n @Override", "score": 0.9388830661773682 }, { "filename": "src/main/java/com/tcoded/folialib/impl/FoliaImplementation.java", "retrieved_chunk": " return new WrappedFoliaTask(\n this.plugin.getServer().getRegionScheduler().runDelayed(\n plugin, location, task -> runnable.run(),\n TimeConverter.toTicks(delay, unit)\n )\n );\n }\n @Override\n public WrappedTask runAtLocationTimer(Location location, Runnable runnable, long delay, long period, TimeUnit unit) {\n return new WrappedFoliaTask(", "score": 0.911758542060852 }, { "filename": "src/main/java/com/tcoded/folialib/impl/FoliaImplementation.java", "retrieved_chunk": " null,\n TimeConverter.toTicks(delay, unit)\n )\n );\n }\n @Override\n public WrappedTask runAtEntityTimer(Entity entity, Runnable runnable, long delay, long period, TimeUnit unit) {\n return new WrappedFoliaTask(\n entity.getScheduler().runAtFixedRate(\n plugin,", "score": 0.9104495048522949 }, { "filename": "src/main/java/com/tcoded/folialib/impl/ServerImplementation.java", "retrieved_chunk": " WrappedTask runTimer(Runnable runnable, long delay, long period, TimeUnit unit);\n /**\n * Folia: Async\n * Paper: Async\n * Spigot: Async\n * @param runnable Task to run\n * @param delay Delay before first execution\n * @param period Delay between executions\n * @param unit Time unit\n * @return WrappedTask instance", "score": 0.9094306826591492 } ]
java
TimeConverter.toTicks(delay, unit), TimeConverter.toTicks(period, unit)) );
/* * Licensed to the Apache Software Foundation (ASF) under one or more * contributor license agreements. See the NOTICE file distributed with * this work for additional information regarding copyright ownership. * The ASF licenses this file to You under the Apache License, Version 2.0 * (the "License"); you may not use this file except in compliance with * the License. You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ package org.tallison.cc.index.extractor; import java.io.ByteArrayInputStream; import java.io.ByteArrayOutputStream; import java.io.IOException; import java.io.InputStream; import java.nio.file.Files; import java.nio.file.Path; import java.nio.file.StandardCopyOption; import java.util.Optional; import java.util.zip.GZIPInputStream; import org.apache.commons.codec.binary.Base32; import org.apache.commons.codec.digest.DigestUtils; import org.apache.commons.io.IOUtils; import org.netpreserve.jwarc.MediaType; import org.netpreserve.jwarc.WarcPayload; import org.netpreserve.jwarc.WarcReader; import org.netpreserve.jwarc.WarcRecord; import org.netpreserve.jwarc.WarcResponse; import org.slf4j.Logger; import org.slf4j.LoggerFactory; import org.tallison.cc.index.CCIndexReaderCounter; import org.tallison.cc.index.CCIndexRecord; import org.tallison.cc.index.io.TargetPathRewriter; import org.apache.tika.exception.TikaConfigException; import org.apache.tika.exception.TikaException; import org.apache.tika.io.TikaInputStream; import org.apache.tika.metadata.Metadata; import org.apache.tika.pipes.FetchEmitTuple; import org.apache.tika.pipes.emitter.EmitKey; import org.apache.tika.pipes.emitter.StreamEmitter; import org.apache.tika.pipes.fetcher.FetchKey; import org.apache.tika.pipes.fetcher.RangeFetcher; public class FileFromCCWarcExtractor { private static Logger LOGGER = LoggerFactory.getLogger(FileFromCCWarcExtractor.class); private static Logger EXTRACTED_LOGGER = LoggerFactory.getLogger("extracted-urls"); private static Logger EXTRACTED_ALL_LOGGER = LoggerFactory.getLogger("extracted-urls-all"); private final StreamEmitter emitter; private final TargetPathRewriter targetPathRewriter; private RangeFetcher fetcher; private final boolean extractTruncated; private Base32 base32 = new Base32(); private final CCIndexReaderCounter ccIndexReaderCounter; public FileFromCCWarcExtractor(ExtractorConfig fetcherConfig, CCIndexReaderCounter ccIndexReaderCounter) throws TikaConfigException { this.emitter = fetcherConfig.newEmitter(); this.fetcher = (RangeFetcher) fetcherConfig.newFetcher(); this.targetPathRewriter = fetcherConfig.getTargetPathRewriter(); this.extractTruncated = fetcherConfig.isExtractTruncated(); this.ccIndexReaderCounter = ccIndexReaderCounter; } public void fetchToPath(CCIndexRecord record) throws InterruptedException { LOGGER
.debug("going to fetch {
} {}->{}", record.getFilename(), record.getOffset(), record.getLength()); FetchEmitTuple t = new FetchEmitTuple(record.getFilename(), new FetchKey("", record.getFilename(), record.getOffset(), record.getOffset() + record.getLength() - 1), new EmitKey()); byte[] warcRecordGZBytes; try { warcRecordGZBytes = fetchWarcBytes(t); } catch (TikaException | IOException e) { LOGGER.warn("couldn't get bytes from cc's warc " + t, e); return; } String id = record.getUrl(); try { parseWarc(id, record, warcRecordGZBytes); } catch (IOException e) { LOGGER.warn("problem parsing warc file", e); } } private void fetchPayload(String id, CCIndexRecord ccIndexRecord, WarcRecord record) throws IOException { if (!((record instanceof WarcResponse) && record.contentType().base().equals(MediaType.HTTP))) { return; } Optional<WarcPayload> payload = ((WarcResponse) record).payload(); if (!payload.isPresent()) { LOGGER.debug("no payload {}", id); ccIndexReaderCounter.getEmptyPayload().incrementAndGet(); return; } if (payload.get().body().size() == 0) { LOGGER.debug("empty payload id={}", id); ccIndexReaderCounter.getEmptyPayload().incrementAndGet(); return; } Path tmp = Files.createTempFile("ccfile-fetcher-", ""); try { Files.copy(payload.get().body().stream(), tmp, StandardCopyOption.REPLACE_EXISTING); String targetDigest = null; String base32Sha1 = ""; try (InputStream is = Files.newInputStream(tmp)) { base32Sha1 = base32.encodeAsString(DigestUtils.sha1(is)); } catch (IOException e) { LOGGER.warn("IOException during digesting: " + tmp.toAbsolutePath()); return; } if (!base32Sha1.equals(ccIndexRecord.getDigest())) { LOGGER.warn("Bad digest for url={} ccindex={} sha1={}", id, ccIndexRecord.getDigest(), base32Sha1); } //TODO: make digest and encoding configurable try (InputStream is = Files.newInputStream(tmp)) { targetDigest = DigestUtils.sha256Hex(is); } catch (IOException e) { LOGGER.warn("IOException during digesting: " + tmp.toAbsolutePath()); return; } long length = -1; try { length = Files.size(tmp); } catch (IOException e) { LOGGER.warn("IOException during digesting: " + tmp.toAbsolutePath()); return; } String targetPath = targetPathRewriter.rewrite(targetDigest); Metadata metadata = new Metadata(); try (InputStream is = TikaInputStream.get(tmp, metadata)) { emitter.emit(targetPath, is, new Metadata()); logSuccess(ccIndexRecord, targetDigest, length, targetPath); } catch (IOException | TikaException e) { LOGGER.warn("problem writing id={}", id, e); } } finally { try { Files.delete(tmp); } catch (IOException e) { LOGGER.warn("can't delete " + tmp.toAbsolutePath(), e); } } } private void logSuccess(CCIndexRecord ccIndexRecord, String targetDigest, long length, String targetPath) { if (extractTruncated) { EXTRACTED_ALL_LOGGER.info("", ccIndexRecord.getUrl(), ccIndexRecord.getNormalizedMime(), ccIndexRecord.getNormalizedMimeDetected(), ccIndexRecord.getFilename(), ccIndexRecord.getOffset(), ccIndexRecord.getLength(), ccIndexRecord.getTruncated(), targetDigest, length, targetPath); } else { //new ObjectArray ? //url,mime_detected,warc_file,warc_offset,warc_length,sha256,length,path EXTRACTED_LOGGER.info("", ccIndexRecord.getUrl(), ccIndexRecord.getNormalizedMime(), ccIndexRecord.getNormalizedMimeDetected(), ccIndexRecord.getFilename(), ccIndexRecord.getOffset(), ccIndexRecord.getLength(), targetDigest, length, targetPath); } } private void parseWarc(String id, CCIndexRecord ccIndexRecord, byte[] warcRecordGZBytes) throws IOException { //need to leave initial inputstream open while parsing warcrecord //can't just parse record and return try (InputStream is = new GZIPInputStream(new ByteArrayInputStream(warcRecordGZBytes))) { try (WarcReader warcreader = new WarcReader(is)) { //should be a single warc per file //return the first for (WarcRecord warcRecord : warcreader) { fetchPayload(id, ccIndexRecord, warcRecord); return; } } } } private byte[] fetchWarcBytes(FetchEmitTuple t) throws TikaException, InterruptedException, IOException { ByteArrayOutputStream bos = new ByteArrayOutputStream(); FetchKey k = t.getFetchKey(); try (InputStream is = fetcher.fetch(k.getFetchKey(), k.getRangeStart(), k.getRangeEnd(), new Metadata())) { IOUtils.copy(is, bos); } return bos.toByteArray(); } }
src/main/java/org/tallison/cc/index/extractor/FileFromCCWarcExtractor.java
tballison-commoncrawl-fetcher-lite-2c106d1
[ { "filename": "src/main/java/org/tallison/cc/index/extractor/CCMimeCounter.java", "retrieved_chunk": "import org.tallison.cc.index.AbstractRecordProcessor;\nimport org.tallison.cc.index.CCIndexReaderCounter;\nimport org.tallison.cc.index.CCIndexRecord;\nimport org.tallison.cc.index.IndexIterator;\nimport org.apache.tika.exception.TikaConfigException;\nimport org.apache.tika.exception.TikaException;\nimport org.apache.tika.io.TikaInputStream;\nimport org.apache.tika.metadata.Metadata;\nimport org.apache.tika.pipes.FetchEmitTuple;\nimport org.apache.tika.pipes.fetcher.Fetcher;", "score": 0.8498815894126892 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/CCFileExtractor.java", "retrieved_chunk": " }\n return INDEX_WORKER_ID;\n }\n private boolean processFile(FetchEmitTuple fetchEmitTuple,\n AbstractRecordProcessor recordProcessor)\n throws InterruptedException {\n long start = System.currentTimeMillis();\n LOGGER.info(\"starting to fetch index gz: {}\",\n fetchEmitTuple.getFetchKey().getFetchKey());\n try (TikaInputStream tis = (TikaInputStream) indexFetcher.fetch(", "score": 0.8493141531944275 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/CCMimeCounter.java", "retrieved_chunk": " private final Fetcher fetcher;\n IndexWorker(ExtractorConfig fetcherConfig, ArrayBlockingQueue<FetchEmitTuple> indexUrls,\n AbstractRecordProcessor recordProcessor) throws TikaException {\n this.indexUrls = indexUrls;\n this.recordProcessor = recordProcessor;\n this.fetcher = fetcherConfig.newFetcher();\n }\n @Override\n public Long call() throws Exception {\n boolean shouldContinue = true;", "score": 0.8472007513046265 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/CCFileExtractorRecordProcessor.java", "retrieved_chunk": " LoggerFactory.getLogger(\"truncated-urls-full\");\n private final ExtractorConfig fetcherConfig;\n private final CCIndexReaderCounter counter;\n private final FileFromCCWarcExtractor fileFromCCWarcFetcher;\n private long reportEvery = 100000;\n public CCFileExtractorRecordProcessor(ExtractorConfig fetcherConfig, CCIndexReaderCounter counter)\n throws TikaConfigException, IOException {\n this.fetcherConfig = fetcherConfig;\n this.counter = counter;\n this.fileFromCCWarcFetcher = new FileFromCCWarcExtractor(fetcherConfig, counter);", "score": 0.8416110277175903 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/ExtractorConfig.java", "retrieved_chunk": " }\n public void setRecordSelector(RecordSelector recordSelector) {\n this.recordSelector = recordSelector;\n }\n public IndexIterator getIndexIterator() {\n return indexIterator;\n }\n public Fetcher newFetcher() throws TikaConfigException {\n if (fetchConfig == null) {\n fetchConfig = new FetchConfig(null, DEFAULT_THROTTLE_SECONDS, null);", "score": 0.8408219814300537 } ]
java
.debug("going to fetch {
package com.tcoded.folialib.impl; import com.tcoded.folialib.FoliaLib; import com.tcoded.folialib.enums.EntityTaskResult; import com.tcoded.folialib.util.TimeConverter; import com.tcoded.folialib.wrapper.WrappedTask; import com.tcoded.folialib.wrapper.task.WrappedBukkitTask; import org.bukkit.Location; import org.bukkit.entity.Entity; import org.bukkit.entity.Player; import org.bukkit.plugin.java.JavaPlugin; import org.bukkit.scheduler.BukkitScheduler; import org.jetbrains.annotations.NotNull; import java.util.UUID; import java.util.concurrent.CompletableFuture; import java.util.concurrent.ExecutionException; import java.util.concurrent.TimeUnit; public class SpigotImplementation implements ServerImplementation { private final JavaPlugin plugin; @SuppressWarnings("deprecation") private final @NotNull BukkitScheduler scheduler; public SpigotImplementation(FoliaLib foliaLib) { this.plugin = foliaLib.getPlugin(); this.scheduler = plugin.getServer().getScheduler(); } @Override public CompletableFuture<Void> runNextTick(Runnable runnable) { CompletableFuture<Void> future = new CompletableFuture<>(); this.scheduler.runTask(plugin, () -> { runnable.run(); future.complete(null); }); return future; } @Override public CompletableFuture<Void> runAsync(Runnable runnable) { CompletableFuture<Void> future = new CompletableFuture<>(); this.scheduler.runTaskAsynchronously(plugin, () -> { runnable.run(); future.complete(null); }); return future; } @Override public WrappedTask runLater(Runnable runnable, long delay, TimeUnit unit) { return new WrappedBukkitTask( this.scheduler.
runTaskLater(plugin, runnable, TimeConverter.toTicks(delay, unit)) );
} @Override public WrappedTask runLaterAsync(Runnable runnable, long delay, TimeUnit unit) { return new WrappedBukkitTask( this.scheduler.runTaskLaterAsynchronously(plugin, runnable, TimeConverter.toTicks(delay, unit)) ); } @Override public WrappedTask runTimer(Runnable runnable, long delay, long period, TimeUnit unit) { return new WrappedBukkitTask( this.scheduler.runTaskTimer( plugin, runnable, TimeConverter.toTicks(delay, unit), TimeConverter.toTicks(period, unit)) ); } @Override public WrappedTask runTimerAsync(Runnable runnable, long delay, long period, TimeUnit unit) { return new WrappedBukkitTask( this.scheduler.runTaskTimerAsynchronously( plugin, runnable, TimeConverter.toTicks(delay, unit), TimeConverter.toTicks(period, unit)) ); } @Override public CompletableFuture<Void> runAtLocation(Location location, Runnable runnable) { CompletableFuture<Void> future = new CompletableFuture<>(); this.scheduler.runTask(plugin, () -> { runnable.run(); future.complete(null); }); return future; } @Override public WrappedTask runAtLocationLater(Location location, Runnable runnable, long delay, TimeUnit unit) { return new WrappedBukkitTask( this.scheduler.runTaskLater(plugin, runnable, TimeConverter.toTicks(delay, unit)) ); } @Override public WrappedTask runAtLocationTimer(Location location, Runnable runnable, long delay, long period, TimeUnit unit) { return new WrappedBukkitTask( this.scheduler.runTaskTimer( plugin, runnable, TimeConverter.toTicks(delay, unit), TimeConverter.toTicks(period, unit)) ); } @Override public CompletableFuture<EntityTaskResult> runAtEntity(Entity entity, Runnable runnable) { CompletableFuture<EntityTaskResult> future = new CompletableFuture<>(); this.scheduler.runTask(plugin, () -> { runnable.run(); future.complete(EntityTaskResult.SUCCESS); }); return future; } @Override public CompletableFuture<EntityTaskResult> runAtEntityWithFallback(Entity entity, Runnable runnable, Runnable fallback) { CompletableFuture<EntityTaskResult> future = new CompletableFuture<>(); this.scheduler.runTask(plugin, () -> { if (entity.isValid()) { runnable.run(); future.complete(EntityTaskResult.SUCCESS); } else { fallback.run(); future.complete(EntityTaskResult.ENTITY_RETIRED); } }); return future; } @Override public WrappedTask runAtEntityLater(Entity entity, Runnable runnable, long delay, TimeUnit unit) { return new WrappedBukkitTask( this.scheduler.runTaskLater(plugin, runnable, TimeConverter.toTicks(delay, unit)) ); } @Override public WrappedTask runAtEntityTimer(Entity entity, Runnable runnable, long delay, long period, TimeUnit unit) { return new WrappedBukkitTask( this.scheduler.runTaskTimer( plugin, runnable, TimeConverter.toTicks(delay, unit), TimeConverter.toTicks(period, unit)) ); } @Override public void cancelTask(WrappedTask task) { task.cancel(); } @Override public void cancelAllTasks() { this.scheduler.cancelTasks(plugin); } @Override public Player getPlayer(String name) { // Already on the main thread if (this.plugin.getServer().isPrimaryThread()) { return this.plugin.getServer().getPlayer(name); } // Not on the main thread, we need to wait until the next tick else { try { return this.scheduler.callSyncMethod(plugin, () -> this.plugin.getServer().getPlayer(name)).get(); } catch (InterruptedException | ExecutionException e) { e.printStackTrace(); } } // Fallback to null return null; } @Override public Player getPlayerExact(String name) { // Already on the main thread if (this.plugin.getServer().isPrimaryThread()) { return this.plugin.getServer().getPlayerExact(name); } // Not on the main thread, we need to wait until the next tick else { try { return this.scheduler.callSyncMethod(plugin, () -> this.plugin.getServer().getPlayerExact(name)).get(); } catch (InterruptedException | ExecutionException e) { e.printStackTrace(); } } // Fallback to null return null; } @SuppressWarnings("DuplicatedCode") @Override public Player getPlayer(UUID uuid) { // Already on the main thread if (this.plugin.getServer().isPrimaryThread()) { return this.plugin.getServer().getPlayer(uuid); } // Not on the main thread, we need to wait until the next tick else { try { return this.scheduler.callSyncMethod(plugin, () -> this.plugin.getServer().getPlayer(uuid)).get(); } catch (InterruptedException | ExecutionException e) { e.printStackTrace(); } } // Fallback to null return null; } @Override public CompletableFuture<Boolean> teleportAsync(Player player, Location location) { CompletableFuture<Boolean> future = new CompletableFuture<>(); this.runAtEntity(player, () -> { if (player.isValid() && player.isOnline()) { player.teleport(location); future.complete(true); } else { future.complete(false); } }); return future; } }
src/main/java/com/tcoded/folialib/impl/SpigotImplementation.java
TechnicallyCoded-FoliaLib-8f9f24f
[ { "filename": "src/main/java/com/tcoded/folialib/impl/FoliaImplementation.java", "retrieved_chunk": " future.complete(null);\n });\n return future;\n }\n @Override\n public WrappedTask runLater(Runnable runnable, long delay, TimeUnit unit) {\n return new WrappedFoliaTask(\n this.globalRegionScheduler.runDelayed(\n plugin, task -> runnable.run(), TimeConverter.toTicks(delay, unit)\n )", "score": 0.9406002759933472 }, { "filename": "src/main/java/com/tcoded/folialib/impl/FoliaImplementation.java", "retrieved_chunk": " );\n }\n @Override\n public WrappedTask runLaterAsync(Runnable runnable, long delay, TimeUnit unit) {\n return new WrappedFoliaTask(\n this.asyncScheduler.runDelayed(\n plugin, task -> runnable.run(), delay, unit\n )\n );\n }", "score": 0.9208428859710693 }, { "filename": "src/main/java/com/tcoded/folialib/impl/FoliaImplementation.java", "retrieved_chunk": " future.complete(EntityTaskResult.SCHEDULER_RETIRED);\n }\n return future;\n }\n @Override\n public WrappedTask runAtEntityLater(Entity entity, Runnable runnable, long delay, TimeUnit unit) {\n return new WrappedFoliaTask(\n entity.getScheduler().runDelayed(\n plugin,\n task -> runnable.run(),", "score": 0.9034953713417053 }, { "filename": "src/main/java/com/tcoded/folialib/impl/FoliaImplementation.java", "retrieved_chunk": " this.plugin.getServer().getRegionScheduler().runAtFixedRate(\n plugin, location, task -> runnable.run(),\n TimeConverter.toTicks(delay, unit),\n TimeConverter.toTicks(period, unit)\n )\n );\n }\n @Override\n public CompletableFuture<EntityTaskResult> runAtEntity(Entity entity, Runnable runnable) {\n CompletableFuture<EntityTaskResult> future = new CompletableFuture<>();", "score": 0.9033270478248596 }, { "filename": "src/main/java/com/tcoded/folialib/impl/FoliaImplementation.java", "retrieved_chunk": " @Override\n public WrappedTask runTimerAsync(Runnable runnable, long delay, long period, TimeUnit unit) {\n return new WrappedFoliaTask(\n this.asyncScheduler.runAtFixedRate(\n plugin, task -> runnable.run(),\n delay, period, unit\n )\n );\n }\n @Override", "score": 0.8957206010818481 } ]
java
runTaskLater(plugin, runnable, TimeConverter.toTicks(delay, unit)) );
/* * Licensed to the Apache Software Foundation (ASF) under one or more * contributor license agreements. See the NOTICE file distributed with * this work for additional information regarding copyright ownership. * The ASF licenses this file to You under the Apache License, Version 2.0 * (the "License"); you may not use this file except in compliance with * the License. You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ package org.tallison.cc.index.extractor; import java.io.ByteArrayInputStream; import java.io.ByteArrayOutputStream; import java.io.IOException; import java.io.InputStream; import java.nio.file.Files; import java.nio.file.Path; import java.nio.file.StandardCopyOption; import java.util.Optional; import java.util.zip.GZIPInputStream; import org.apache.commons.codec.binary.Base32; import org.apache.commons.codec.digest.DigestUtils; import org.apache.commons.io.IOUtils; import org.netpreserve.jwarc.MediaType; import org.netpreserve.jwarc.WarcPayload; import org.netpreserve.jwarc.WarcReader; import org.netpreserve.jwarc.WarcRecord; import org.netpreserve.jwarc.WarcResponse; import org.slf4j.Logger; import org.slf4j.LoggerFactory; import org.tallison.cc.index.CCIndexReaderCounter; import org.tallison.cc.index.CCIndexRecord; import org.tallison.cc.index.io.TargetPathRewriter; import org.apache.tika.exception.TikaConfigException; import org.apache.tika.exception.TikaException; import org.apache.tika.io.TikaInputStream; import org.apache.tika.metadata.Metadata; import org.apache.tika.pipes.FetchEmitTuple; import org.apache.tika.pipes.emitter.EmitKey; import org.apache.tika.pipes.emitter.StreamEmitter; import org.apache.tika.pipes.fetcher.FetchKey; import org.apache.tika.pipes.fetcher.RangeFetcher; public class FileFromCCWarcExtractor { private static Logger LOGGER = LoggerFactory.getLogger(FileFromCCWarcExtractor.class); private static Logger EXTRACTED_LOGGER = LoggerFactory.getLogger("extracted-urls"); private static Logger EXTRACTED_ALL_LOGGER = LoggerFactory.getLogger("extracted-urls-all"); private final StreamEmitter emitter; private final TargetPathRewriter targetPathRewriter; private RangeFetcher fetcher; private final boolean extractTruncated; private Base32 base32 = new Base32(); private final CCIndexReaderCounter ccIndexReaderCounter; public FileFromCCWarcExtractor(ExtractorConfig fetcherConfig, CCIndexReaderCounter ccIndexReaderCounter) throws TikaConfigException { this.emitter = fetcherConfig.newEmitter(); this.fetcher = (RangeFetcher) fetcherConfig.newFetcher(); this.targetPathRewriter = fetcherConfig.getTargetPathRewriter(); this.extractTruncated = fetcherConfig.isExtractTruncated(); this.ccIndexReaderCounter = ccIndexReaderCounter; } public void fetchToPath(CCIndexRecord record) throws InterruptedException { LOGGER.debug("going to fetch {} {}->{}", record.getFilename(), record.getOffset(),
record.getLength());
FetchEmitTuple t = new FetchEmitTuple(record.getFilename(), new FetchKey("", record.getFilename(), record.getOffset(), record.getOffset() + record.getLength() - 1), new EmitKey()); byte[] warcRecordGZBytes; try { warcRecordGZBytes = fetchWarcBytes(t); } catch (TikaException | IOException e) { LOGGER.warn("couldn't get bytes from cc's warc " + t, e); return; } String id = record.getUrl(); try { parseWarc(id, record, warcRecordGZBytes); } catch (IOException e) { LOGGER.warn("problem parsing warc file", e); } } private void fetchPayload(String id, CCIndexRecord ccIndexRecord, WarcRecord record) throws IOException { if (!((record instanceof WarcResponse) && record.contentType().base().equals(MediaType.HTTP))) { return; } Optional<WarcPayload> payload = ((WarcResponse) record).payload(); if (!payload.isPresent()) { LOGGER.debug("no payload {}", id); ccIndexReaderCounter.getEmptyPayload().incrementAndGet(); return; } if (payload.get().body().size() == 0) { LOGGER.debug("empty payload id={}", id); ccIndexReaderCounter.getEmptyPayload().incrementAndGet(); return; } Path tmp = Files.createTempFile("ccfile-fetcher-", ""); try { Files.copy(payload.get().body().stream(), tmp, StandardCopyOption.REPLACE_EXISTING); String targetDigest = null; String base32Sha1 = ""; try (InputStream is = Files.newInputStream(tmp)) { base32Sha1 = base32.encodeAsString(DigestUtils.sha1(is)); } catch (IOException e) { LOGGER.warn("IOException during digesting: " + tmp.toAbsolutePath()); return; } if (!base32Sha1.equals(ccIndexRecord.getDigest())) { LOGGER.warn("Bad digest for url={} ccindex={} sha1={}", id, ccIndexRecord.getDigest(), base32Sha1); } //TODO: make digest and encoding configurable try (InputStream is = Files.newInputStream(tmp)) { targetDigest = DigestUtils.sha256Hex(is); } catch (IOException e) { LOGGER.warn("IOException during digesting: " + tmp.toAbsolutePath()); return; } long length = -1; try { length = Files.size(tmp); } catch (IOException e) { LOGGER.warn("IOException during digesting: " + tmp.toAbsolutePath()); return; } String targetPath = targetPathRewriter.rewrite(targetDigest); Metadata metadata = new Metadata(); try (InputStream is = TikaInputStream.get(tmp, metadata)) { emitter.emit(targetPath, is, new Metadata()); logSuccess(ccIndexRecord, targetDigest, length, targetPath); } catch (IOException | TikaException e) { LOGGER.warn("problem writing id={}", id, e); } } finally { try { Files.delete(tmp); } catch (IOException e) { LOGGER.warn("can't delete " + tmp.toAbsolutePath(), e); } } } private void logSuccess(CCIndexRecord ccIndexRecord, String targetDigest, long length, String targetPath) { if (extractTruncated) { EXTRACTED_ALL_LOGGER.info("", ccIndexRecord.getUrl(), ccIndexRecord.getNormalizedMime(), ccIndexRecord.getNormalizedMimeDetected(), ccIndexRecord.getFilename(), ccIndexRecord.getOffset(), ccIndexRecord.getLength(), ccIndexRecord.getTruncated(), targetDigest, length, targetPath); } else { //new ObjectArray ? //url,mime_detected,warc_file,warc_offset,warc_length,sha256,length,path EXTRACTED_LOGGER.info("", ccIndexRecord.getUrl(), ccIndexRecord.getNormalizedMime(), ccIndexRecord.getNormalizedMimeDetected(), ccIndexRecord.getFilename(), ccIndexRecord.getOffset(), ccIndexRecord.getLength(), targetDigest, length, targetPath); } } private void parseWarc(String id, CCIndexRecord ccIndexRecord, byte[] warcRecordGZBytes) throws IOException { //need to leave initial inputstream open while parsing warcrecord //can't just parse record and return try (InputStream is = new GZIPInputStream(new ByteArrayInputStream(warcRecordGZBytes))) { try (WarcReader warcreader = new WarcReader(is)) { //should be a single warc per file //return the first for (WarcRecord warcRecord : warcreader) { fetchPayload(id, ccIndexRecord, warcRecord); return; } } } } private byte[] fetchWarcBytes(FetchEmitTuple t) throws TikaException, InterruptedException, IOException { ByteArrayOutputStream bos = new ByteArrayOutputStream(); FetchKey k = t.getFetchKey(); try (InputStream is = fetcher.fetch(k.getFetchKey(), k.getRangeStart(), k.getRangeEnd(), new Metadata())) { IOUtils.copy(is, bos); } return bos.toByteArray(); } }
src/main/java/org/tallison/cc/index/extractor/FileFromCCWarcExtractor.java
tballison-commoncrawl-fetcher-lite-2c106d1
[ { "filename": "src/main/java/org/tallison/cc/index/extractor/CCMimeCounter.java", "retrieved_chunk": "import org.tallison.cc.index.AbstractRecordProcessor;\nimport org.tallison.cc.index.CCIndexReaderCounter;\nimport org.tallison.cc.index.CCIndexRecord;\nimport org.tallison.cc.index.IndexIterator;\nimport org.apache.tika.exception.TikaConfigException;\nimport org.apache.tika.exception.TikaException;\nimport org.apache.tika.io.TikaInputStream;\nimport org.apache.tika.metadata.Metadata;\nimport org.apache.tika.pipes.FetchEmitTuple;\nimport org.apache.tika.pipes.fetcher.Fetcher;", "score": 0.8673235774040222 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/CCMimeCounter.java", "retrieved_chunk": " private final Fetcher fetcher;\n IndexWorker(ExtractorConfig fetcherConfig, ArrayBlockingQueue<FetchEmitTuple> indexUrls,\n AbstractRecordProcessor recordProcessor) throws TikaException {\n this.indexUrls = indexUrls;\n this.recordProcessor = recordProcessor;\n this.fetcher = fetcherConfig.newFetcher();\n }\n @Override\n public Long call() throws Exception {\n boolean shouldContinue = true;", "score": 0.8626571297645569 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/CCFileExtractorRecordProcessor.java", "retrieved_chunk": " LoggerFactory.getLogger(\"truncated-urls-full\");\n private final ExtractorConfig fetcherConfig;\n private final CCIndexReaderCounter counter;\n private final FileFromCCWarcExtractor fileFromCCWarcFetcher;\n private long reportEvery = 100000;\n public CCFileExtractorRecordProcessor(ExtractorConfig fetcherConfig, CCIndexReaderCounter counter)\n throws TikaConfigException, IOException {\n this.fetcherConfig = fetcherConfig;\n this.counter = counter;\n this.fileFromCCWarcFetcher = new FileFromCCWarcExtractor(fetcherConfig, counter);", "score": 0.8583172559738159 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/CCFileExtractor.java", "retrieved_chunk": " ExtractorConfig fetcherConfig =\n new ObjectMapper().readValue(new File(args[0]), ExtractorConfig.class);\n execute(fetcherConfig);\n }\n private static void execute(ExtractorConfig fetcherConfig) throws TikaException {\n ArrayBlockingQueue<FetchEmitTuple> indexPathsList = new ArrayBlockingQueue<>(1000);\n //IndexPathsReader reads a file containing a list of cc-index.paths files\n //and writes the literal gz files (cc-index/collections/CC-MAIN-2023-06/indexes/cdx-00000.gz)\n //to indexPathsList\n //IndexWorker reads a single index.gz file at a time and processes each record", "score": 0.8555663824081421 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/ExtractorConfig.java", "retrieved_chunk": " }\n public void setRecordSelector(RecordSelector recordSelector) {\n this.recordSelector = recordSelector;\n }\n public IndexIterator getIndexIterator() {\n return indexIterator;\n }\n public Fetcher newFetcher() throws TikaConfigException {\n if (fetchConfig == null) {\n fetchConfig = new FetchConfig(null, DEFAULT_THROTTLE_SECONDS, null);", "score": 0.855474054813385 } ]
java
record.getLength());
/* * Licensed to the Apache Software Foundation (ASF) under one or more * contributor license agreements. See the NOTICE file distributed with * this work for additional information regarding copyright ownership. * The ASF licenses this file to You under the Apache License, Version 2.0 * (the "License"); you may not use this file except in compliance with * the License. You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ package org.tallison.cc.index.extractor; import java.io.ByteArrayInputStream; import java.io.ByteArrayOutputStream; import java.io.IOException; import java.io.InputStream; import java.nio.file.Files; import java.nio.file.Path; import java.nio.file.StandardCopyOption; import java.util.Optional; import java.util.zip.GZIPInputStream; import org.apache.commons.codec.binary.Base32; import org.apache.commons.codec.digest.DigestUtils; import org.apache.commons.io.IOUtils; import org.netpreserve.jwarc.MediaType; import org.netpreserve.jwarc.WarcPayload; import org.netpreserve.jwarc.WarcReader; import org.netpreserve.jwarc.WarcRecord; import org.netpreserve.jwarc.WarcResponse; import org.slf4j.Logger; import org.slf4j.LoggerFactory; import org.tallison.cc.index.CCIndexReaderCounter; import org.tallison.cc.index.CCIndexRecord; import org.tallison.cc.index.io.TargetPathRewriter; import org.apache.tika.exception.TikaConfigException; import org.apache.tika.exception.TikaException; import org.apache.tika.io.TikaInputStream; import org.apache.tika.metadata.Metadata; import org.apache.tika.pipes.FetchEmitTuple; import org.apache.tika.pipes.emitter.EmitKey; import org.apache.tika.pipes.emitter.StreamEmitter; import org.apache.tika.pipes.fetcher.FetchKey; import org.apache.tika.pipes.fetcher.RangeFetcher; public class FileFromCCWarcExtractor { private static Logger LOGGER = LoggerFactory.getLogger(FileFromCCWarcExtractor.class); private static Logger EXTRACTED_LOGGER = LoggerFactory.getLogger("extracted-urls"); private static Logger EXTRACTED_ALL_LOGGER = LoggerFactory.getLogger("extracted-urls-all"); private final StreamEmitter emitter; private final TargetPathRewriter targetPathRewriter; private RangeFetcher fetcher; private final boolean extractTruncated; private Base32 base32 = new Base32(); private final CCIndexReaderCounter ccIndexReaderCounter; public FileFromCCWarcExtractor(ExtractorConfig fetcherConfig, CCIndexReaderCounter ccIndexReaderCounter) throws TikaConfigException { this.emitter = fetcherConfig.newEmitter(); this.fetcher = (RangeFetcher) fetcherConfig.newFetcher(); this.targetPathRewriter = fetcherConfig.getTargetPathRewriter(); this.extractTruncated = fetcherConfig.isExtractTruncated(); this.ccIndexReaderCounter = ccIndexReaderCounter; } public void fetchToPath(CCIndexRecord record) throws InterruptedException { LOGGER.debug("going to fetch {} {}->{}", record.getFilename(), record.getOffset(), record.getLength()); FetchEmitTuple t = new FetchEmitTuple(record.getFilename(), new FetchKey("", record.getFilename(), record.getOffset(),
record.getOffset() + record.getLength() - 1), new EmitKey());
byte[] warcRecordGZBytes; try { warcRecordGZBytes = fetchWarcBytes(t); } catch (TikaException | IOException e) { LOGGER.warn("couldn't get bytes from cc's warc " + t, e); return; } String id = record.getUrl(); try { parseWarc(id, record, warcRecordGZBytes); } catch (IOException e) { LOGGER.warn("problem parsing warc file", e); } } private void fetchPayload(String id, CCIndexRecord ccIndexRecord, WarcRecord record) throws IOException { if (!((record instanceof WarcResponse) && record.contentType().base().equals(MediaType.HTTP))) { return; } Optional<WarcPayload> payload = ((WarcResponse) record).payload(); if (!payload.isPresent()) { LOGGER.debug("no payload {}", id); ccIndexReaderCounter.getEmptyPayload().incrementAndGet(); return; } if (payload.get().body().size() == 0) { LOGGER.debug("empty payload id={}", id); ccIndexReaderCounter.getEmptyPayload().incrementAndGet(); return; } Path tmp = Files.createTempFile("ccfile-fetcher-", ""); try { Files.copy(payload.get().body().stream(), tmp, StandardCopyOption.REPLACE_EXISTING); String targetDigest = null; String base32Sha1 = ""; try (InputStream is = Files.newInputStream(tmp)) { base32Sha1 = base32.encodeAsString(DigestUtils.sha1(is)); } catch (IOException e) { LOGGER.warn("IOException during digesting: " + tmp.toAbsolutePath()); return; } if (!base32Sha1.equals(ccIndexRecord.getDigest())) { LOGGER.warn("Bad digest for url={} ccindex={} sha1={}", id, ccIndexRecord.getDigest(), base32Sha1); } //TODO: make digest and encoding configurable try (InputStream is = Files.newInputStream(tmp)) { targetDigest = DigestUtils.sha256Hex(is); } catch (IOException e) { LOGGER.warn("IOException during digesting: " + tmp.toAbsolutePath()); return; } long length = -1; try { length = Files.size(tmp); } catch (IOException e) { LOGGER.warn("IOException during digesting: " + tmp.toAbsolutePath()); return; } String targetPath = targetPathRewriter.rewrite(targetDigest); Metadata metadata = new Metadata(); try (InputStream is = TikaInputStream.get(tmp, metadata)) { emitter.emit(targetPath, is, new Metadata()); logSuccess(ccIndexRecord, targetDigest, length, targetPath); } catch (IOException | TikaException e) { LOGGER.warn("problem writing id={}", id, e); } } finally { try { Files.delete(tmp); } catch (IOException e) { LOGGER.warn("can't delete " + tmp.toAbsolutePath(), e); } } } private void logSuccess(CCIndexRecord ccIndexRecord, String targetDigest, long length, String targetPath) { if (extractTruncated) { EXTRACTED_ALL_LOGGER.info("", ccIndexRecord.getUrl(), ccIndexRecord.getNormalizedMime(), ccIndexRecord.getNormalizedMimeDetected(), ccIndexRecord.getFilename(), ccIndexRecord.getOffset(), ccIndexRecord.getLength(), ccIndexRecord.getTruncated(), targetDigest, length, targetPath); } else { //new ObjectArray ? //url,mime_detected,warc_file,warc_offset,warc_length,sha256,length,path EXTRACTED_LOGGER.info("", ccIndexRecord.getUrl(), ccIndexRecord.getNormalizedMime(), ccIndexRecord.getNormalizedMimeDetected(), ccIndexRecord.getFilename(), ccIndexRecord.getOffset(), ccIndexRecord.getLength(), targetDigest, length, targetPath); } } private void parseWarc(String id, CCIndexRecord ccIndexRecord, byte[] warcRecordGZBytes) throws IOException { //need to leave initial inputstream open while parsing warcrecord //can't just parse record and return try (InputStream is = new GZIPInputStream(new ByteArrayInputStream(warcRecordGZBytes))) { try (WarcReader warcreader = new WarcReader(is)) { //should be a single warc per file //return the first for (WarcRecord warcRecord : warcreader) { fetchPayload(id, ccIndexRecord, warcRecord); return; } } } } private byte[] fetchWarcBytes(FetchEmitTuple t) throws TikaException, InterruptedException, IOException { ByteArrayOutputStream bos = new ByteArrayOutputStream(); FetchKey k = t.getFetchKey(); try (InputStream is = fetcher.fetch(k.getFetchKey(), k.getRangeStart(), k.getRangeEnd(), new Metadata())) { IOUtils.copy(is, bos); } return bos.toByteArray(); } }
src/main/java/org/tallison/cc/index/extractor/FileFromCCWarcExtractor.java
tballison-commoncrawl-fetcher-lite-2c106d1
[ { "filename": "src/main/java/org/tallison/cc/index/extractor/CCFileExtractor.java", "retrieved_chunk": " }\n return INDEX_WORKER_ID;\n }\n private boolean processFile(FetchEmitTuple fetchEmitTuple,\n AbstractRecordProcessor recordProcessor)\n throws InterruptedException {\n long start = System.currentTimeMillis();\n LOGGER.info(\"starting to fetch index gz: {}\",\n fetchEmitTuple.getFetchKey().getFetchKey());\n try (TikaInputStream tis = (TikaInputStream) indexFetcher.fetch(", "score": 0.8548794984817505 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/CCFileExtractor.java", "retrieved_chunk": " this.indexUrls = indexUrls;\n this.recordProcessor = recordProcessor;\n this.indexFetcher = fetcherConfig.newIndexFetcher();\n }\n @Override\n public Long call() throws Exception {\n boolean shouldContinue = true;\n while (shouldContinue) {\n FetchEmitTuple indexUrl = indexUrls.poll(120, TimeUnit.MINUTES);\n if (indexUrl == null) {", "score": 0.8539709448814392 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/CCMimeCounter.java", "retrieved_chunk": " private final Fetcher fetcher;\n IndexWorker(ExtractorConfig fetcherConfig, ArrayBlockingQueue<FetchEmitTuple> indexUrls,\n AbstractRecordProcessor recordProcessor) throws TikaException {\n this.indexUrls = indexUrls;\n this.recordProcessor = recordProcessor;\n this.fetcher = fetcherConfig.newFetcher();\n }\n @Override\n public Long call() throws Exception {\n boolean shouldContinue = true;", "score": 0.8523125648498535 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/CCMimeCounter.java", "retrieved_chunk": "import org.tallison.cc.index.AbstractRecordProcessor;\nimport org.tallison.cc.index.CCIndexReaderCounter;\nimport org.tallison.cc.index.CCIndexRecord;\nimport org.tallison.cc.index.IndexIterator;\nimport org.apache.tika.exception.TikaConfigException;\nimport org.apache.tika.exception.TikaException;\nimport org.apache.tika.io.TikaInputStream;\nimport org.apache.tika.metadata.Metadata;\nimport org.apache.tika.pipes.FetchEmitTuple;\nimport org.apache.tika.pipes.fetcher.Fetcher;", "score": 0.8487301468849182 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/CCFileExtractorRecordProcessor.java", "retrieved_chunk": " LoggerFactory.getLogger(\"truncated-urls-full\");\n private final ExtractorConfig fetcherConfig;\n private final CCIndexReaderCounter counter;\n private final FileFromCCWarcExtractor fileFromCCWarcFetcher;\n private long reportEvery = 100000;\n public CCFileExtractorRecordProcessor(ExtractorConfig fetcherConfig, CCIndexReaderCounter counter)\n throws TikaConfigException, IOException {\n this.fetcherConfig = fetcherConfig;\n this.counter = counter;\n this.fileFromCCWarcFetcher = new FileFromCCWarcExtractor(fetcherConfig, counter);", "score": 0.8433579206466675 } ]
java
record.getOffset() + record.getLength() - 1), new EmitKey());
/* * Licensed to the Apache Software Foundation (ASF) under one or more * contributor license agreements. See the NOTICE file distributed with * this work for additional information regarding copyright ownership. * The ASF licenses this file to You under the Apache License, Version 2.0 * (the "License"); you may not use this file except in compliance with * the License. You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ package org.tallison.cc.index.extractor; import java.io.ByteArrayInputStream; import java.io.ByteArrayOutputStream; import java.io.IOException; import java.io.InputStream; import java.nio.file.Files; import java.nio.file.Path; import java.nio.file.StandardCopyOption; import java.util.Optional; import java.util.zip.GZIPInputStream; import org.apache.commons.codec.binary.Base32; import org.apache.commons.codec.digest.DigestUtils; import org.apache.commons.io.IOUtils; import org.netpreserve.jwarc.MediaType; import org.netpreserve.jwarc.WarcPayload; import org.netpreserve.jwarc.WarcReader; import org.netpreserve.jwarc.WarcRecord; import org.netpreserve.jwarc.WarcResponse; import org.slf4j.Logger; import org.slf4j.LoggerFactory; import org.tallison.cc.index.CCIndexReaderCounter; import org.tallison.cc.index.CCIndexRecord; import org.tallison.cc.index.io.TargetPathRewriter; import org.apache.tika.exception.TikaConfigException; import org.apache.tika.exception.TikaException; import org.apache.tika.io.TikaInputStream; import org.apache.tika.metadata.Metadata; import org.apache.tika.pipes.FetchEmitTuple; import org.apache.tika.pipes.emitter.EmitKey; import org.apache.tika.pipes.emitter.StreamEmitter; import org.apache.tika.pipes.fetcher.FetchKey; import org.apache.tika.pipes.fetcher.RangeFetcher; public class FileFromCCWarcExtractor { private static Logger LOGGER = LoggerFactory.getLogger(FileFromCCWarcExtractor.class); private static Logger EXTRACTED_LOGGER = LoggerFactory.getLogger("extracted-urls"); private static Logger EXTRACTED_ALL_LOGGER = LoggerFactory.getLogger("extracted-urls-all"); private final StreamEmitter emitter; private final TargetPathRewriter targetPathRewriter; private RangeFetcher fetcher; private final boolean extractTruncated; private Base32 base32 = new Base32(); private final CCIndexReaderCounter ccIndexReaderCounter; public FileFromCCWarcExtractor(ExtractorConfig fetcherConfig, CCIndexReaderCounter ccIndexReaderCounter) throws TikaConfigException { this.emitter = fetcherConfig.newEmitter(); this.fetcher = (RangeFetcher) fetcherConfig.newFetcher(); this.targetPathRewriter = fetcherConfig.getTargetPathRewriter(); this.extractTruncated = fetcherConfig.isExtractTruncated(); this.ccIndexReaderCounter = ccIndexReaderCounter; } public void fetchToPath(CCIndexRecord record) throws InterruptedException { LOGGER.debug("going to fetch {} {}->{}", record.getFilename(), record.getOffset(), record.getLength()); FetchEmitTuple t = new FetchEmitTuple(record.getFilename(), new FetchKey("", record.getFilename(), record.getOffset(), record.getOffset() + record.getLength() - 1), new EmitKey()); byte[] warcRecordGZBytes; try { warcRecordGZBytes = fetchWarcBytes(t); } catch (TikaException | IOException e) { LOGGER.warn("couldn't get bytes from cc's warc " + t, e); return; } String id = record.getUrl(); try { parseWarc(id, record, warcRecordGZBytes); } catch (IOException e) { LOGGER.warn("problem parsing warc file", e); } } private void fetchPayload(String id, CCIndexRecord ccIndexRecord, WarcRecord record) throws IOException { if (!((record instanceof WarcResponse) && record.contentType().base().equals(MediaType.HTTP))) { return; } Optional<WarcPayload> payload = ((WarcResponse) record).payload(); if (!payload.isPresent()) { LOGGER.debug("no payload {}", id); ccIndexReaderCounter.getEmptyPayload().incrementAndGet(); return; } if (payload.get().body().size() == 0) { LOGGER.debug("empty payload id={}", id); ccIndexReaderCounter.getEmptyPayload().incrementAndGet(); return; } Path tmp = Files.createTempFile("ccfile-fetcher-", ""); try { Files.copy(payload.get().body().stream(), tmp, StandardCopyOption.REPLACE_EXISTING); String targetDigest = null; String base32Sha1 = ""; try (InputStream is = Files.newInputStream(tmp)) { base32Sha1 = base32.encodeAsString(DigestUtils.sha1(is)); } catch (IOException e) { LOGGER.warn("IOException during digesting: " + tmp.toAbsolutePath()); return; } if (!base32Sha1
.equals(ccIndexRecord.getDigest())) {
LOGGER.warn("Bad digest for url={} ccindex={} sha1={}", id, ccIndexRecord.getDigest(), base32Sha1); } //TODO: make digest and encoding configurable try (InputStream is = Files.newInputStream(tmp)) { targetDigest = DigestUtils.sha256Hex(is); } catch (IOException e) { LOGGER.warn("IOException during digesting: " + tmp.toAbsolutePath()); return; } long length = -1; try { length = Files.size(tmp); } catch (IOException e) { LOGGER.warn("IOException during digesting: " + tmp.toAbsolutePath()); return; } String targetPath = targetPathRewriter.rewrite(targetDigest); Metadata metadata = new Metadata(); try (InputStream is = TikaInputStream.get(tmp, metadata)) { emitter.emit(targetPath, is, new Metadata()); logSuccess(ccIndexRecord, targetDigest, length, targetPath); } catch (IOException | TikaException e) { LOGGER.warn("problem writing id={}", id, e); } } finally { try { Files.delete(tmp); } catch (IOException e) { LOGGER.warn("can't delete " + tmp.toAbsolutePath(), e); } } } private void logSuccess(CCIndexRecord ccIndexRecord, String targetDigest, long length, String targetPath) { if (extractTruncated) { EXTRACTED_ALL_LOGGER.info("", ccIndexRecord.getUrl(), ccIndexRecord.getNormalizedMime(), ccIndexRecord.getNormalizedMimeDetected(), ccIndexRecord.getFilename(), ccIndexRecord.getOffset(), ccIndexRecord.getLength(), ccIndexRecord.getTruncated(), targetDigest, length, targetPath); } else { //new ObjectArray ? //url,mime_detected,warc_file,warc_offset,warc_length,sha256,length,path EXTRACTED_LOGGER.info("", ccIndexRecord.getUrl(), ccIndexRecord.getNormalizedMime(), ccIndexRecord.getNormalizedMimeDetected(), ccIndexRecord.getFilename(), ccIndexRecord.getOffset(), ccIndexRecord.getLength(), targetDigest, length, targetPath); } } private void parseWarc(String id, CCIndexRecord ccIndexRecord, byte[] warcRecordGZBytes) throws IOException { //need to leave initial inputstream open while parsing warcrecord //can't just parse record and return try (InputStream is = new GZIPInputStream(new ByteArrayInputStream(warcRecordGZBytes))) { try (WarcReader warcreader = new WarcReader(is)) { //should be a single warc per file //return the first for (WarcRecord warcRecord : warcreader) { fetchPayload(id, ccIndexRecord, warcRecord); return; } } } } private byte[] fetchWarcBytes(FetchEmitTuple t) throws TikaException, InterruptedException, IOException { ByteArrayOutputStream bos = new ByteArrayOutputStream(); FetchKey k = t.getFetchKey(); try (InputStream is = fetcher.fetch(k.getFetchKey(), k.getRangeStart(), k.getRangeEnd(), new Metadata())) { IOUtils.copy(is, bos); } return bos.toByteArray(); } }
src/main/java/org/tallison/cc/index/extractor/FileFromCCWarcExtractor.java
tballison-commoncrawl-fetcher-lite-2c106d1
[ { "filename": "src/test/java/org/tallison/cc/index/selector/IndexRecordSelectorTest.java", "retrieved_chunk": "import java.nio.file.Files;\nimport java.nio.file.Path;\nimport java.nio.file.Paths;\nimport java.util.Optional;\nimport java.util.zip.GZIPInputStream;\nimport com.fasterxml.jackson.databind.ObjectMapper;\nimport org.junit.jupiter.api.Disabled;\nimport org.junit.jupiter.api.Test;\nimport org.tallison.cc.index.CCIndexRecord;\nimport org.apache.tika.utils.StringUtils;", "score": 0.8261247873306274 }, { "filename": "src/test/java/org/tallison/cc/index/selector/IndexRecordSelectorTest.java", "retrieved_chunk": " public void testIndexFile() throws Exception {\n Path p = Paths.get(\"/Users/allison/data/cc/index-work/cdx-00000.gz\");\n try (BufferedReader r = new BufferedReader(\n new InputStreamReader(new GZIPInputStream(Files.newInputStream(p)),\n StandardCharsets.UTF_8))) {\n String line = r.readLine();\n while (line != null) {\n Optional<CCIndexRecord> record = CCIndexRecord.parseRecord(line);\n if (record.isPresent()) {\n CCIndexRecord indexRecord = record.get();", "score": 0.8203513622283936 }, { "filename": "src/test/java/org/tallison/cc/index/selector/IndexRecordSelectorTest.java", "retrieved_chunk": " if (!indexRecord.getMime().equals(indexRecord.getMimeDetected())) {\n System.out.println(line);\n }\n if (!StringUtils.isBlank(indexRecord.getTruncated())) {\n }\n }\n line = r.readLine();\n }\n }\n }", "score": 0.7885791063308716 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/CCFileExtractor.java", "retrieved_chunk": "import org.apache.tika.utils.StringUtils;\n/**\n * This is a lighter class that doesn't rely on a database\n * to extract files from CC and log a list of truncated urls.\n */\npublic class CCFileExtractor {\n private static final Long INDEX_WORKER_ID = 1l;\n private static final Long INDEX_READER_ID = 2l;\n private static final Logger LOGGER = LoggerFactory.getLogger(CCFileExtractor.class);\n public static void main(String[] args) throws Exception {", "score": 0.7870038151741028 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/CCFileExtractorRecordProcessor.java", "retrieved_chunk": " String url = r.getUrl();\n TRUNCATED_URLS_LOGGER.info(\"\", url);\n //url,mime,mime_detected,warc_file,warc_offset,warc_length,truncated\n TRUNCATED_URLS_FULL_LOGGER.info(\"\", url,\n r.getNormalizedMime(), r.getNormalizedMimeDetected(), r.getFilename(),\n r.getOffset(), r.getLength(), r.getTruncated());\n return true;\n }\n }\n private void fetchBytes(CCIndexRecord r) throws InterruptedException {", "score": 0.7858055233955383 } ]
java
.equals(ccIndexRecord.getDigest())) {
/* * Licensed to the Apache Software Foundation (ASF) under one or more * contributor license agreements. See the NOTICE file distributed with * this work for additional information regarding copyright ownership. * The ASF licenses this file to You under the Apache License, Version 2.0 * (the "License"); you may not use this file except in compliance with * the License. You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ package org.tallison.cc.index.extractor; import java.io.ByteArrayInputStream; import java.io.ByteArrayOutputStream; import java.io.IOException; import java.io.InputStream; import java.nio.file.Files; import java.nio.file.Path; import java.nio.file.StandardCopyOption; import java.util.Optional; import java.util.zip.GZIPInputStream; import org.apache.commons.codec.binary.Base32; import org.apache.commons.codec.digest.DigestUtils; import org.apache.commons.io.IOUtils; import org.netpreserve.jwarc.MediaType; import org.netpreserve.jwarc.WarcPayload; import org.netpreserve.jwarc.WarcReader; import org.netpreserve.jwarc.WarcRecord; import org.netpreserve.jwarc.WarcResponse; import org.slf4j.Logger; import org.slf4j.LoggerFactory; import org.tallison.cc.index.CCIndexReaderCounter; import org.tallison.cc.index.CCIndexRecord; import org.tallison.cc.index.io.TargetPathRewriter; import org.apache.tika.exception.TikaConfigException; import org.apache.tika.exception.TikaException; import org.apache.tika.io.TikaInputStream; import org.apache.tika.metadata.Metadata; import org.apache.tika.pipes.FetchEmitTuple; import org.apache.tika.pipes.emitter.EmitKey; import org.apache.tika.pipes.emitter.StreamEmitter; import org.apache.tika.pipes.fetcher.FetchKey; import org.apache.tika.pipes.fetcher.RangeFetcher; public class FileFromCCWarcExtractor { private static Logger LOGGER = LoggerFactory.getLogger(FileFromCCWarcExtractor.class); private static Logger EXTRACTED_LOGGER = LoggerFactory.getLogger("extracted-urls"); private static Logger EXTRACTED_ALL_LOGGER = LoggerFactory.getLogger("extracted-urls-all"); private final StreamEmitter emitter; private final TargetPathRewriter targetPathRewriter; private RangeFetcher fetcher; private final boolean extractTruncated; private Base32 base32 = new Base32(); private final CCIndexReaderCounter ccIndexReaderCounter; public FileFromCCWarcExtractor(ExtractorConfig fetcherConfig, CCIndexReaderCounter ccIndexReaderCounter) throws TikaConfigException { this.emitter = fetcherConfig.newEmitter(); this.fetcher = (RangeFetcher) fetcherConfig.newFetcher(); this.targetPathRewriter = fetcherConfig.getTargetPathRewriter(); this.extractTruncated = fetcherConfig.isExtractTruncated(); this.ccIndexReaderCounter = ccIndexReaderCounter; } public void fetchToPath(CCIndexRecord record) throws InterruptedException { LOGGER.debug("going to fetch {} {}->{}", record.getFilename(), record.getOffset(), record.getLength()); FetchEmitTuple t = new FetchEmitTuple(record.getFilename(), new FetchKey("", record.getFilename(), record.getOffset(), record
.getOffset() + record.getLength() - 1), new EmitKey());
byte[] warcRecordGZBytes; try { warcRecordGZBytes = fetchWarcBytes(t); } catch (TikaException | IOException e) { LOGGER.warn("couldn't get bytes from cc's warc " + t, e); return; } String id = record.getUrl(); try { parseWarc(id, record, warcRecordGZBytes); } catch (IOException e) { LOGGER.warn("problem parsing warc file", e); } } private void fetchPayload(String id, CCIndexRecord ccIndexRecord, WarcRecord record) throws IOException { if (!((record instanceof WarcResponse) && record.contentType().base().equals(MediaType.HTTP))) { return; } Optional<WarcPayload> payload = ((WarcResponse) record).payload(); if (!payload.isPresent()) { LOGGER.debug("no payload {}", id); ccIndexReaderCounter.getEmptyPayload().incrementAndGet(); return; } if (payload.get().body().size() == 0) { LOGGER.debug("empty payload id={}", id); ccIndexReaderCounter.getEmptyPayload().incrementAndGet(); return; } Path tmp = Files.createTempFile("ccfile-fetcher-", ""); try { Files.copy(payload.get().body().stream(), tmp, StandardCopyOption.REPLACE_EXISTING); String targetDigest = null; String base32Sha1 = ""; try (InputStream is = Files.newInputStream(tmp)) { base32Sha1 = base32.encodeAsString(DigestUtils.sha1(is)); } catch (IOException e) { LOGGER.warn("IOException during digesting: " + tmp.toAbsolutePath()); return; } if (!base32Sha1.equals(ccIndexRecord.getDigest())) { LOGGER.warn("Bad digest for url={} ccindex={} sha1={}", id, ccIndexRecord.getDigest(), base32Sha1); } //TODO: make digest and encoding configurable try (InputStream is = Files.newInputStream(tmp)) { targetDigest = DigestUtils.sha256Hex(is); } catch (IOException e) { LOGGER.warn("IOException during digesting: " + tmp.toAbsolutePath()); return; } long length = -1; try { length = Files.size(tmp); } catch (IOException e) { LOGGER.warn("IOException during digesting: " + tmp.toAbsolutePath()); return; } String targetPath = targetPathRewriter.rewrite(targetDigest); Metadata metadata = new Metadata(); try (InputStream is = TikaInputStream.get(tmp, metadata)) { emitter.emit(targetPath, is, new Metadata()); logSuccess(ccIndexRecord, targetDigest, length, targetPath); } catch (IOException | TikaException e) { LOGGER.warn("problem writing id={}", id, e); } } finally { try { Files.delete(tmp); } catch (IOException e) { LOGGER.warn("can't delete " + tmp.toAbsolutePath(), e); } } } private void logSuccess(CCIndexRecord ccIndexRecord, String targetDigest, long length, String targetPath) { if (extractTruncated) { EXTRACTED_ALL_LOGGER.info("", ccIndexRecord.getUrl(), ccIndexRecord.getNormalizedMime(), ccIndexRecord.getNormalizedMimeDetected(), ccIndexRecord.getFilename(), ccIndexRecord.getOffset(), ccIndexRecord.getLength(), ccIndexRecord.getTruncated(), targetDigest, length, targetPath); } else { //new ObjectArray ? //url,mime_detected,warc_file,warc_offset,warc_length,sha256,length,path EXTRACTED_LOGGER.info("", ccIndexRecord.getUrl(), ccIndexRecord.getNormalizedMime(), ccIndexRecord.getNormalizedMimeDetected(), ccIndexRecord.getFilename(), ccIndexRecord.getOffset(), ccIndexRecord.getLength(), targetDigest, length, targetPath); } } private void parseWarc(String id, CCIndexRecord ccIndexRecord, byte[] warcRecordGZBytes) throws IOException { //need to leave initial inputstream open while parsing warcrecord //can't just parse record and return try (InputStream is = new GZIPInputStream(new ByteArrayInputStream(warcRecordGZBytes))) { try (WarcReader warcreader = new WarcReader(is)) { //should be a single warc per file //return the first for (WarcRecord warcRecord : warcreader) { fetchPayload(id, ccIndexRecord, warcRecord); return; } } } } private byte[] fetchWarcBytes(FetchEmitTuple t) throws TikaException, InterruptedException, IOException { ByteArrayOutputStream bos = new ByteArrayOutputStream(); FetchKey k = t.getFetchKey(); try (InputStream is = fetcher.fetch(k.getFetchKey(), k.getRangeStart(), k.getRangeEnd(), new Metadata())) { IOUtils.copy(is, bos); } return bos.toByteArray(); } }
src/main/java/org/tallison/cc/index/extractor/FileFromCCWarcExtractor.java
tballison-commoncrawl-fetcher-lite-2c106d1
[ { "filename": "src/main/java/org/tallison/cc/index/extractor/CCFileExtractor.java", "retrieved_chunk": " }\n return INDEX_WORKER_ID;\n }\n private boolean processFile(FetchEmitTuple fetchEmitTuple,\n AbstractRecordProcessor recordProcessor)\n throws InterruptedException {\n long start = System.currentTimeMillis();\n LOGGER.info(\"starting to fetch index gz: {}\",\n fetchEmitTuple.getFetchKey().getFetchKey());\n try (TikaInputStream tis = (TikaInputStream) indexFetcher.fetch(", "score": 0.8575912714004517 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/CCFileExtractor.java", "retrieved_chunk": " this.indexUrls = indexUrls;\n this.recordProcessor = recordProcessor;\n this.indexFetcher = fetcherConfig.newIndexFetcher();\n }\n @Override\n public Long call() throws Exception {\n boolean shouldContinue = true;\n while (shouldContinue) {\n FetchEmitTuple indexUrl = indexUrls.poll(120, TimeUnit.MINUTES);\n if (indexUrl == null) {", "score": 0.8533511757850647 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/CCMimeCounter.java", "retrieved_chunk": "import org.tallison.cc.index.AbstractRecordProcessor;\nimport org.tallison.cc.index.CCIndexReaderCounter;\nimport org.tallison.cc.index.CCIndexRecord;\nimport org.tallison.cc.index.IndexIterator;\nimport org.apache.tika.exception.TikaConfigException;\nimport org.apache.tika.exception.TikaException;\nimport org.apache.tika.io.TikaInputStream;\nimport org.apache.tika.metadata.Metadata;\nimport org.apache.tika.pipes.FetchEmitTuple;\nimport org.apache.tika.pipes.fetcher.Fetcher;", "score": 0.8532038927078247 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/CCFileExtractorRecordProcessor.java", "retrieved_chunk": " //problem already logged\n return true;\n }\n CCIndexRecord r = record.get();\n if (!fetcherConfig.getRecordSelector().select(r)) {\n return true;\n }\n //if truncated, count appropriately and test for limits\n if (!StringUtils.isBlank(r.getTruncated())) {\n long truncated = counter.getTruncated().incrementAndGet();", "score": 0.8518154621124268 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/CCMimeCounter.java", "retrieved_chunk": " private final Fetcher fetcher;\n IndexWorker(ExtractorConfig fetcherConfig, ArrayBlockingQueue<FetchEmitTuple> indexUrls,\n AbstractRecordProcessor recordProcessor) throws TikaException {\n this.indexUrls = indexUrls;\n this.recordProcessor = recordProcessor;\n this.fetcher = fetcherConfig.newFetcher();\n }\n @Override\n public Long call() throws Exception {\n boolean shouldContinue = true;", "score": 0.8500478267669678 } ]
java
.getOffset() + record.getLength() - 1), new EmitKey());
/* * Licensed to the Apache Software Foundation (ASF) under one or more * contributor license agreements. See the NOTICE file distributed with * this work for additional information regarding copyright ownership. * The ASF licenses this file to You under the Apache License, Version 2.0 * (the "License"); you may not use this file except in compliance with * the License. You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ package org.tallison.cc.index.extractor; import java.io.ByteArrayInputStream; import java.io.ByteArrayOutputStream; import java.io.IOException; import java.io.InputStream; import java.nio.file.Files; import java.nio.file.Path; import java.nio.file.StandardCopyOption; import java.util.Optional; import java.util.zip.GZIPInputStream; import org.apache.commons.codec.binary.Base32; import org.apache.commons.codec.digest.DigestUtils; import org.apache.commons.io.IOUtils; import org.netpreserve.jwarc.MediaType; import org.netpreserve.jwarc.WarcPayload; import org.netpreserve.jwarc.WarcReader; import org.netpreserve.jwarc.WarcRecord; import org.netpreserve.jwarc.WarcResponse; import org.slf4j.Logger; import org.slf4j.LoggerFactory; import org.tallison.cc.index.CCIndexReaderCounter; import org.tallison.cc.index.CCIndexRecord; import org.tallison.cc.index.io.TargetPathRewriter; import org.apache.tika.exception.TikaConfigException; import org.apache.tika.exception.TikaException; import org.apache.tika.io.TikaInputStream; import org.apache.tika.metadata.Metadata; import org.apache.tika.pipes.FetchEmitTuple; import org.apache.tika.pipes.emitter.EmitKey; import org.apache.tika.pipes.emitter.StreamEmitter; import org.apache.tika.pipes.fetcher.FetchKey; import org.apache.tika.pipes.fetcher.RangeFetcher; public class FileFromCCWarcExtractor { private static Logger LOGGER = LoggerFactory.getLogger(FileFromCCWarcExtractor.class); private static Logger EXTRACTED_LOGGER = LoggerFactory.getLogger("extracted-urls"); private static Logger EXTRACTED_ALL_LOGGER = LoggerFactory.getLogger("extracted-urls-all"); private final StreamEmitter emitter; private final TargetPathRewriter targetPathRewriter; private RangeFetcher fetcher; private final boolean extractTruncated; private Base32 base32 = new Base32(); private final CCIndexReaderCounter ccIndexReaderCounter; public FileFromCCWarcExtractor(ExtractorConfig fetcherConfig, CCIndexReaderCounter ccIndexReaderCounter) throws TikaConfigException { this.emitter = fetcherConfig.newEmitter(); this.fetcher = (RangeFetcher) fetcherConfig.newFetcher(); this.targetPathRewriter = fetcherConfig.getTargetPathRewriter(); this.extractTruncated = fetcherConfig.isExtractTruncated(); this.ccIndexReaderCounter = ccIndexReaderCounter; } public void fetchToPath(CCIndexRecord record) throws InterruptedException { LOGGER.debug("going to fetch {} {}->{}", record.getFilename(), record.getOffset(), record.getLength()); FetchEmitTuple t = new FetchEmitTuple(record.getFilename(), new FetchKey("", record.getFilename(), record.getOffset(), record.getOffset() + record.getLength() - 1), new EmitKey()); byte[] warcRecordGZBytes; try { warcRecordGZBytes = fetchWarcBytes(t); } catch (TikaException | IOException e) { LOGGER.warn("couldn't get bytes from cc's warc " + t, e); return; } String id = record.getUrl(); try { parseWarc(id, record, warcRecordGZBytes); } catch (IOException e) { LOGGER.warn("problem parsing warc file", e); } } private void fetchPayload(String id, CCIndexRecord ccIndexRecord, WarcRecord record) throws IOException { if (!((record instanceof WarcResponse) && record.contentType().base().equals(MediaType.HTTP))) { return; } Optional<WarcPayload> payload = ((WarcResponse) record).payload(); if (!payload.isPresent()) { LOGGER.debug("no payload {}", id); ccIndexReaderCounter.getEmptyPayload().incrementAndGet(); return; } if (payload.get().body().size() == 0) { LOGGER.debug("empty payload id={}", id); ccIndexReaderCounter.getEmptyPayload().incrementAndGet(); return; } Path tmp = Files.createTempFile("ccfile-fetcher-", ""); try { Files.copy(payload.get().body().stream(), tmp, StandardCopyOption.REPLACE_EXISTING); String targetDigest = null; String base32Sha1 = ""; try (InputStream is = Files.newInputStream(tmp)) { base32Sha1 = base32.encodeAsString(DigestUtils.sha1(is)); } catch (IOException e) { LOGGER.warn("IOException during digesting: " + tmp.toAbsolutePath()); return; } if (!base32Sha1.equals(ccIndexRecord.getDigest())) { LOGGER.warn("Bad digest for url={} ccindex={} sha1={}", id, ccIndexRecord.getDigest(), base32Sha1); } //TODO: make digest and encoding configurable try (InputStream is = Files.newInputStream(tmp)) { targetDigest = DigestUtils.sha256Hex(is); } catch (IOException e) { LOGGER.warn("IOException during digesting: " + tmp.toAbsolutePath()); return; } long length = -1; try { length = Files.size(tmp); } catch (IOException e) { LOGGER.warn("IOException during digesting: " + tmp.toAbsolutePath()); return; } String targetPath = targetPathRewriter.rewrite(targetDigest); Metadata metadata = new Metadata(); try (InputStream is = TikaInputStream.get(tmp, metadata)) { emitter.emit(targetPath, is, new Metadata()); logSuccess(ccIndexRecord, targetDigest, length, targetPath); } catch (IOException | TikaException e) { LOGGER.warn("problem writing id={}", id, e); } } finally { try { Files.delete(tmp); } catch (IOException e) { LOGGER.warn("can't delete " + tmp.toAbsolutePath(), e); } } } private void logSuccess(CCIndexRecord ccIndexRecord, String targetDigest, long length, String targetPath) { if (extractTruncated) { EXTRACTED_ALL_LOGGER.info("", ccIndexRecord.getUrl(), ccIndexRecord.getNormalizedMime(), ccIndexRecord.getNormalizedMimeDetected(), ccIndexRecord.getFilename(),
ccIndexRecord.getOffset(), ccIndexRecord.getLength(), ccIndexRecord.getTruncated(), targetDigest, length, targetPath);
} else { //new ObjectArray ? //url,mime_detected,warc_file,warc_offset,warc_length,sha256,length,path EXTRACTED_LOGGER.info("", ccIndexRecord.getUrl(), ccIndexRecord.getNormalizedMime(), ccIndexRecord.getNormalizedMimeDetected(), ccIndexRecord.getFilename(), ccIndexRecord.getOffset(), ccIndexRecord.getLength(), targetDigest, length, targetPath); } } private void parseWarc(String id, CCIndexRecord ccIndexRecord, byte[] warcRecordGZBytes) throws IOException { //need to leave initial inputstream open while parsing warcrecord //can't just parse record and return try (InputStream is = new GZIPInputStream(new ByteArrayInputStream(warcRecordGZBytes))) { try (WarcReader warcreader = new WarcReader(is)) { //should be a single warc per file //return the first for (WarcRecord warcRecord : warcreader) { fetchPayload(id, ccIndexRecord, warcRecord); return; } } } } private byte[] fetchWarcBytes(FetchEmitTuple t) throws TikaException, InterruptedException, IOException { ByteArrayOutputStream bos = new ByteArrayOutputStream(); FetchKey k = t.getFetchKey(); try (InputStream is = fetcher.fetch(k.getFetchKey(), k.getRangeStart(), k.getRangeEnd(), new Metadata())) { IOUtils.copy(is, bos); } return bos.toByteArray(); } }
src/main/java/org/tallison/cc/index/extractor/FileFromCCWarcExtractor.java
tballison-commoncrawl-fetcher-lite-2c106d1
[ { "filename": "src/main/java/org/tallison/cc/index/extractor/CCFileExtractorRecordProcessor.java", "retrieved_chunk": "import org.slf4j.LoggerFactory;\nimport org.tallison.cc.index.AbstractRecordProcessor;\nimport org.tallison.cc.index.CCIndexReaderCounter;\nimport org.tallison.cc.index.CCIndexRecord;\nimport org.apache.tika.exception.TikaConfigException;\nimport org.apache.tika.utils.StringUtils;\npublic class CCFileExtractorRecordProcessor extends AbstractRecordProcessor {\n private static Logger LOGGER = LoggerFactory.getLogger(CCFileExtractorRecordProcessor.class);\n private static Logger TRUNCATED_URLS_LOGGER = LoggerFactory.getLogger(\"truncated-urls\");\n private static Logger TRUNCATED_URLS_FULL_LOGGER =", "score": 0.8359473943710327 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/CCFileExtractor.java", "retrieved_chunk": "import org.apache.tika.utils.StringUtils;\n/**\n * This is a lighter class that doesn't rely on a database\n * to extract files from CC and log a list of truncated urls.\n */\npublic class CCFileExtractor {\n private static final Long INDEX_WORKER_ID = 1l;\n private static final Long INDEX_READER_ID = 2l;\n private static final Logger LOGGER = LoggerFactory.getLogger(CCFileExtractor.class);\n public static void main(String[] args) throws Exception {", "score": 0.8191252946853638 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/CCFileExtractorRecordProcessor.java", "retrieved_chunk": " String url = r.getUrl();\n TRUNCATED_URLS_LOGGER.info(\"\", url);\n //url,mime,mime_detected,warc_file,warc_offset,warc_length,truncated\n TRUNCATED_URLS_FULL_LOGGER.info(\"\", url,\n r.getNormalizedMime(), r.getNormalizedMimeDetected(), r.getFilename(),\n r.getOffset(), r.getLength(), r.getTruncated());\n return true;\n }\n }\n private void fetchBytes(CCIndexRecord r) throws InterruptedException {", "score": 0.8133255243301392 }, { "filename": "src/main/java/org/tallison/cc/index/CCIndexRecord.java", "retrieved_chunk": "import com.fasterxml.jackson.databind.PropertyNamingStrategies;\nimport com.fasterxml.jackson.databind.annotation.JsonNaming;\nimport org.slf4j.Logger;\nimport org.slf4j.LoggerFactory;\n@JsonNaming(PropertyNamingStrategies.KebabCaseStrategy.class)\npublic class CCIndexRecord {\n private static Pattern INT_PATTERN = Pattern.compile(\"^\\\\d+$\");\n private static Logger LOGGER = LoggerFactory.getLogger(CCIndexRecord.class);\n private static ObjectMapper OBJECT_MAPPER = new ObjectMapper();\n private String url;", "score": 0.8049837946891785 }, { "filename": "src/test/java/org/tallison/cc/index/selector/IndexRecordSelectorTest.java", "retrieved_chunk": " public void testIndexFile() throws Exception {\n Path p = Paths.get(\"/Users/allison/data/cc/index-work/cdx-00000.gz\");\n try (BufferedReader r = new BufferedReader(\n new InputStreamReader(new GZIPInputStream(Files.newInputStream(p)),\n StandardCharsets.UTF_8))) {\n String line = r.readLine();\n while (line != null) {\n Optional<CCIndexRecord> record = CCIndexRecord.parseRecord(line);\n if (record.isPresent()) {\n CCIndexRecord indexRecord = record.get();", "score": 0.8046241402626038 } ]
java
ccIndexRecord.getOffset(), ccIndexRecord.getLength(), ccIndexRecord.getTruncated(), targetDigest, length, targetPath);
/* * Licensed to the Apache Software Foundation (ASF) under one or more * contributor license agreements. See the NOTICE file distributed with * this work for additional information regarding copyright ownership. * The ASF licenses this file to You under the Apache License, Version 2.0 * (the "License"); you may not use this file except in compliance with * the License. You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ package org.tallison.cc.index.extractor; import java.io.BufferedInputStream; import java.io.BufferedReader; import java.io.BufferedWriter; import java.io.File; import java.io.IOException; import java.io.InputStream; import java.io.InputStreamReader; import java.nio.charset.StandardCharsets; import java.nio.file.Files; import java.nio.file.Paths; import java.util.ArrayList; import java.util.Collections; import java.util.HashMap; import java.util.List; import java.util.Locale; import java.util.Map; import java.util.Optional; import java.util.concurrent.ArrayBlockingQueue; import java.util.concurrent.Callable; import java.util.concurrent.ExecutionException; import java.util.concurrent.ExecutorCompletionService; import java.util.concurrent.ExecutorService; import java.util.concurrent.Executors; import java.util.concurrent.Future; import java.util.concurrent.TimeUnit; import java.util.concurrent.TimeoutException; import java.util.zip.GZIPInputStream; import com.fasterxml.jackson.databind.ObjectMapper; import org.apache.commons.csv.CSVFormat; import org.apache.commons.csv.CSVPrinter; import org.apache.commons.lang3.mutable.MutableLong; import org.slf4j.Logger; import org.slf4j.LoggerFactory; import org.tallison.cc.index.AbstractRecordProcessor; import org.tallison.cc.index.CCIndexReaderCounter; import org.tallison.cc.index.CCIndexRecord; import org.tallison.cc.index.IndexIterator; import org.apache.tika.exception.TikaConfigException; import org.apache.tika.exception.TikaException; import org.apache.tika.io.TikaInputStream; import org.apache.tika.metadata.Metadata; import org.apache.tika.pipes.FetchEmitTuple; import org.apache.tika.pipes.fetcher.Fetcher; import org.apache.tika.pipes.pipesiterator.CallablePipesIterator; import org.apache.tika.pipes.pipesiterator.PipesIterator; import org.apache.tika.utils.StringUtils; /** * This counts mime_detected. Use a regular file selector to include * only urls that had a 200, e.g. */ public class CCMimeCounter { private static final Long INDEX_WORKER_ID = 1l; private static final Long INDEX_READER_ID = 2l; private static final Logger LOGGER = LoggerFactory.getLogger(CCMimeCounter.class); public static void main(String[] args) throws Exception { ExtractorConfig fetcherConfig = new ObjectMapper().readValue(new File(args[0]), ExtractorConfig.class); execute(fetcherConfig); } private static void execute(ExtractorConfig fetcherConfig) throws IOException, TikaException { ArrayBlockingQueue<FetchEmitTuple> indexPathsList = new ArrayBlockingQueue<>(1000); //IndexPathsReader reads a file containing a list of cc-index.paths files //and writes the literal gz files (cc-index/collections/CC-MAIN-2023-06/indexes/cdx-00000.gz) //to indexPathsList //IndexWorker reads a single index.gz file at a time and processes each record //It fetches non truncated files and logs truncated files int totalThreads = fetcherConfig.getNumThreads() + 1; ExecutorService executorService = Executors.newFixedThreadPool(totalThreads); ExecutorCompletionService<Long> executorCompletionService = new ExecutorCompletionService<>(executorService); IndexIterator indexIterator = fetcherConfig.getIndexIterator(); indexIterator.initialize(Collections.EMPTY_MAP); executorCompletionService.submit(new CallablePipesIterator(indexIterator, indexPathsList)); CCIndexReaderCounter counter = new CCIndexReaderCounter(); int finishedWorkers = 0; List<DetectedMimeCounter> detectedMimeCounters = new ArrayList<>(); try { for (int i = 0; i < fetcherConfig.getNumThreads(); i++) { DetectedMimeCounter processor = new DetectedMimeCounter(fetcherConfig, counter); detectedMimeCounters.add(processor); executorCompletionService.submit( new IndexWorker(fetcherConfig, indexPathsList, processor)); } while (finishedWorkers < totalThreads) { //blocking Future<Long> future = executorCompletionService.take(); if (future != null) { Long f = future.get(); LOGGER.debug("completed {}", f); if (f.equals(INDEX_WORKER_ID)) { finishedWorkers++; } else if (f.equals(INDEX_READER_ID)) { LOGGER.info("Index paths reader successfully completed"); } } } } catch (TikaConfigException e) { LOGGER.error("main loop exception", e); throw new RuntimeException(e); } catch (ExecutionException e) { LOGGER.error("main loop exception", e); throw new RuntimeException(e); } catch (InterruptedException e) { LOGGER.warn("main loop interrupted exception", e); throw new RuntimeException(e); } finally { executorService.shutdown(); executorService.shutdownNow(); } summarize(detectedMimeCounters); } private static void summarize(List<DetectedMimeCounter> detectedMimeCounters) throws IOException { Map<String, Long> total = new HashMap<>(); Map<String, Long> truncated = new HashMap<>(); Map<String, Long> nonTruncated = new HashMap<>(); for (DetectedMimeCounter c : detectedMimeCounters) { update(c.totalCounts, total); update(c.truncatedCounts, truncated); } calcNonTruncated(truncated, total, nonTruncated); report("total", total); report("truncated", truncated); report("non-truncated", nonTruncated); } private static void calcNonTruncated(Map<String, Long> truncated, Map<String, Long> total, Map<String, Long> nonTruncated) { for (Map.Entry<String, Long> e : total.entrySet()) { Long val = e.getValue(); Long t = truncated.getOrDefault(e.getKey(), 0l); val -= t; nonTruncated.put(e.getKey(), val); } } private static void report(String name, Map<String, Long> m) throws IOException { try (BufferedWriter writer = Files.newBufferedWriter(Paths.get(name + ".csv"), StandardCharsets.UTF_8)) { try (CSVPrinter printer = new CSVPrinter(writer, CSVFormat.EXCEL)) { printer.printRecord("mime", "count"); m.entrySet().stream().sorted(Collections.reverseOrder(Map.Entry.comparingByValue())) .forEach(e -> { try { printer.printRecord(e.getKey(), e.getValue()); } catch (IOException ex) { throw new RuntimeException(ex); } }); } } } private static void update(Map<String, MutableLong> from, Map<String, Long> to) { for (Map.Entry<String, MutableLong> e : from.entrySet()) { Long cnt = to.get(e.getKey()); if (cnt == null) { cnt = 0l; } cnt += e.getValue().getValue(); to.put(e.getKey(), cnt); } } private static class IndexWorker implements Callable<Long> { private final ArrayBlockingQueue<FetchEmitTuple> indexUrls; private final AbstractRecordProcessor recordProcessor; private final Fetcher fetcher; IndexWorker(ExtractorConfig fetcherConfig, ArrayBlockingQueue<FetchEmitTuple> indexUrls, AbstractRecordProcessor recordProcessor) throws TikaException { this.indexUrls = indexUrls; this.recordProcessor = recordProcessor; this.fetcher = fetcherConfig.newFetcher(); } @Override public Long call() throws Exception { boolean shouldContinue = true; while (shouldContinue) { FetchEmitTuple indexUrl = indexUrls.poll(120, TimeUnit.MINUTES); if (indexUrl == null) { throw new TimeoutException("waited 120 minutes for a new record"); } if (indexUrl == PipesIterator.COMPLETED_SEMAPHORE) { recordProcessor.close(); //can hang forever indexUrls.put(PipesIterator.COMPLETED_SEMAPHORE); return INDEX_WORKER_ID; } shouldContinue = processFile(indexUrl, recordProcessor); } return INDEX_WORKER_ID; } private boolean processFile(FetchEmitTuple fetchEmitTuple, AbstractRecordProcessor recordProcessor) throws InterruptedException { long start = System.currentTimeMillis(); LOGGER.info("starting to fetch index gz: {}", fetchEmitTuple.getFetchKey().getFetchKey()); try (TikaInputStream tis = (TikaInputStream) fetcher.fetch( fetchEmitTuple.getFetchKey().getFetchKey(), new Metadata())) { try (InputStream is = new BufferedInputStream(new GZIPInputStream(tis))) { try (BufferedReader reader = new BufferedReader( new InputStreamReader(is, StandardCharsets.UTF_8))) { String line = reader.readLine(); int lines = 0; long elapsed = System.currentTimeMillis() - start; LOGGER.info("Finished fetching {} bytes in {} ms for index gz: {}", String.format(Locale.US, "%,d", tis.getLength()), String.format(Locale.US, "%,d", elapsed), fetchEmitTuple.getFetchKey().getFetchKey()); while (line != null) { LOGGER.trace("about to add a line"); if (StringUtils.isBlank(line)) { line = reader.readLine(); continue; } try { boolean shouldContinue = recordProcessor.process(line); if (!shouldContinue) { return shouldContinue; } } catch (IOException e) { LOGGER.warn("bad json: " + line); } lines++; line = reader.readLine(); } } } } catch (TikaException | IOException e) { LOGGER.error( "failed while processing " + fetchEmitTuple.getFetchKey().getFetchKey(), e); } long elapsed = System.currentTimeMillis() - start; LOGGER.info("finished processing index gz in ({}) ms: {}", String.format(Locale.US, "%,d", elapsed), fetchEmitTuple.getFetchKey().getFetchKey()); return true; } } private static class DetectedMimeCounter extends AbstractRecordProcessor { private final ExtractorConfig fetcherConfig; private final CCIndexReaderCounter counter; private final Map<String, MutableLong> totalCounts = new HashMap<>(); private final Map<String, MutableLong> truncatedCounts = new HashMap<>(); public DetectedMimeCounter(ExtractorConfig fetcherConfig, CCIndexReaderCounter counter) { this.fetcherConfig = fetcherConfig; this.counter = counter; } @Override public boolean process(String json) throws IOException, InterruptedException { long totalRead = counter.getRecordsRead().incrementAndGet(); if (totalRead % 1000000 == 0) { LOGGER.info("processed: {}", counter); } if (fetcherConfig.getMaxRecords() > -1 && totalRead >= fetcherConfig.getMaxRecords()) { LOGGER.info("hit max read"); return false; } //check for hit max //return false; Optional<CCIndexRecord> record = CCIndexRecord.parseRecord(json); if (record.isEmpty()) { //problem already logged return true; } CCIndexRecord r = record.get(); if (!fetcherConfig.getRecordSelector().select(r)) { return true; } increment(
totalCounts, r.getNormalizedMimeDetected());
if (!StringUtils.isBlank(r.getTruncated())) { long truncated = counter.getTruncated().incrementAndGet(); if (fetcherConfig.getMaxFilesTruncated() > -1 && truncated >= fetcherConfig.getMaxFilesTruncated()) { LOGGER.info("hit max truncated files"); return false; } increment(truncatedCounts, r.getNormalizedMimeDetected()); return true; } return true; } private void increment(Map<String, MutableLong> m, String k) { MutableLong cnt = m.get(k); if (cnt == null) { cnt = new MutableLong(1); m.put(k, cnt); return; } else { cnt.increment(); } } @Override public void close() throws IOException { } } }
src/main/java/org/tallison/cc/index/extractor/CCMimeCounter.java
tballison-commoncrawl-fetcher-lite-2c106d1
[ { "filename": "src/main/java/org/tallison/cc/index/extractor/CCFileExtractorRecordProcessor.java", "retrieved_chunk": " //problem already logged\n return true;\n }\n CCIndexRecord r = record.get();\n if (!fetcherConfig.getRecordSelector().select(r)) {\n return true;\n }\n //if truncated, count appropriately and test for limits\n if (!StringUtils.isBlank(r.getTruncated())) {\n long truncated = counter.getTruncated().incrementAndGet();", "score": 0.9153097867965698 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/CCFileExtractorRecordProcessor.java", "retrieved_chunk": " LOGGER.info(\"processed: {}\", counter);\n }\n if (fetcherConfig.getMaxRecords() > -1 && totalRead >= fetcherConfig.getMaxRecords()) {\n LOGGER.info(\"hit max read\");\n return false;\n }\n //check for hit max\n //return false;\n Optional<CCIndexRecord> record = CCIndexRecord.parseRecord(json);\n if (record.isEmpty()) {", "score": 0.8559080958366394 }, { "filename": "src/main/java/org/tallison/cc/index/selector/RecordSelector.java", "retrieved_chunk": " if (clause.select(val)) {\n return true;\n }\n }\n }\n return false;\n }\n private String getStringValue(String key, CCIndexRecord record) {\n switch (key) {\n case \"mime_detected\":", "score": 0.8475366234779358 }, { "filename": "src/main/java/org/tallison/cc/index/selector/RecordSelector.java", "retrieved_chunk": " return record.getHost();\n case \"digest\":\n return record.getDigest();\n default:\n throw new IllegalArgumentException(\"Don't yet support key \" + key);\n }\n }\n private static class AcceptAllRecords extends RecordSelector {\n @Override\n public boolean select(CCIndexRecord record) {", "score": 0.8379336595535278 }, { "filename": "src/main/java/org/tallison/cc/index/selector/RecordSelector.java", "retrieved_chunk": " return record.getMimeDetected();\n case \"truncated\":\n return record.getTruncated();\n case \"mime\":\n return record.getMime();\n case \"status\":\n return Integer.toString(record.getStatus());\n case \"url\":\n return record.getUrl();\n case \"host\":", "score": 0.8342307806015015 } ]
java
totalCounts, r.getNormalizedMimeDetected());
/* * Licensed to the Apache Software Foundation (ASF) under one or more * contributor license agreements. See the NOTICE file distributed with * this work for additional information regarding copyright ownership. * The ASF licenses this file to You under the Apache License, Version 2.0 * (the "License"); you may not use this file except in compliance with * the License. You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ package org.tallison.cc.index.extractor; import java.io.ByteArrayInputStream; import java.io.ByteArrayOutputStream; import java.io.IOException; import java.io.InputStream; import java.nio.file.Files; import java.nio.file.Path; import java.nio.file.StandardCopyOption; import java.util.Optional; import java.util.zip.GZIPInputStream; import org.apache.commons.codec.binary.Base32; import org.apache.commons.codec.digest.DigestUtils; import org.apache.commons.io.IOUtils; import org.netpreserve.jwarc.MediaType; import org.netpreserve.jwarc.WarcPayload; import org.netpreserve.jwarc.WarcReader; import org.netpreserve.jwarc.WarcRecord; import org.netpreserve.jwarc.WarcResponse; import org.slf4j.Logger; import org.slf4j.LoggerFactory; import org.tallison.cc.index.CCIndexReaderCounter; import org.tallison.cc.index.CCIndexRecord; import org.tallison.cc.index.io.TargetPathRewriter; import org.apache.tika.exception.TikaConfigException; import org.apache.tika.exception.TikaException; import org.apache.tika.io.TikaInputStream; import org.apache.tika.metadata.Metadata; import org.apache.tika.pipes.FetchEmitTuple; import org.apache.tika.pipes.emitter.EmitKey; import org.apache.tika.pipes.emitter.StreamEmitter; import org.apache.tika.pipes.fetcher.FetchKey; import org.apache.tika.pipes.fetcher.RangeFetcher; public class FileFromCCWarcExtractor { private static Logger LOGGER = LoggerFactory.getLogger(FileFromCCWarcExtractor.class); private static Logger EXTRACTED_LOGGER = LoggerFactory.getLogger("extracted-urls"); private static Logger EXTRACTED_ALL_LOGGER = LoggerFactory.getLogger("extracted-urls-all"); private final StreamEmitter emitter; private final TargetPathRewriter targetPathRewriter; private RangeFetcher fetcher; private final boolean extractTruncated; private Base32 base32 = new Base32(); private final CCIndexReaderCounter ccIndexReaderCounter; public FileFromCCWarcExtractor(ExtractorConfig fetcherConfig, CCIndexReaderCounter ccIndexReaderCounter) throws TikaConfigException { this.emitter = fetcherConfig.newEmitter(); this.fetcher = (RangeFetcher) fetcherConfig.newFetcher(); this.targetPathRewriter = fetcherConfig.getTargetPathRewriter(); this.extractTruncated = fetcherConfig.isExtractTruncated(); this.ccIndexReaderCounter = ccIndexReaderCounter; } public void fetchToPath(CCIndexRecord record) throws InterruptedException { LOGGER.debug("going to fetch {} {}->{}", record.getFilename(), record.getOffset(), record.getLength()); FetchEmitTuple t = new FetchEmitTuple(record.getFilename(), new FetchKey("", record.getFilename(), record.getOffset(), record.getOffset() + record.getLength() - 1), new EmitKey()); byte[] warcRecordGZBytes; try { warcRecordGZBytes = fetchWarcBytes(t); } catch (TikaException | IOException e) { LOGGER.warn("couldn't get bytes from cc's warc " + t, e); return; } String id = record.getUrl(); try { parseWarc(id, record, warcRecordGZBytes); } catch (IOException e) { LOGGER.warn("problem parsing warc file", e); } } private void fetchPayload(String id, CCIndexRecord ccIndexRecord, WarcRecord record) throws IOException { if (!((record instanceof WarcResponse) && record.contentType().base().equals(MediaType.HTTP))) { return; } Optional<WarcPayload> payload = ((WarcResponse) record).payload(); if (!payload.isPresent()) { LOGGER.debug("no payload {}", id); ccIndexReaderCounter.getEmptyPayload().incrementAndGet(); return; } if (payload.get().body().size() == 0) { LOGGER.debug("empty payload id={}", id); ccIndexReaderCounter.getEmptyPayload().incrementAndGet(); return; } Path tmp = Files.createTempFile("ccfile-fetcher-", ""); try { Files.copy(payload.get().body().stream(), tmp, StandardCopyOption.REPLACE_EXISTING); String targetDigest = null; String base32Sha1 = ""; try (InputStream is = Files.newInputStream(tmp)) { base32Sha1 = base32.encodeAsString(DigestUtils.sha1(is)); } catch (IOException e) { LOGGER.warn("IOException during digesting: " + tmp.toAbsolutePath()); return; } if (!base32Sha1.equals(ccIndexRecord.getDigest())) { LOGGER.warn("Bad digest for url={} ccindex={} sha1={}", id, ccIndexRecord.getDigest(), base32Sha1); } //TODO: make digest and encoding configurable try (InputStream is = Files.newInputStream(tmp)) { targetDigest = DigestUtils.sha256Hex(is); } catch (IOException e) { LOGGER.warn("IOException during digesting: " + tmp.toAbsolutePath()); return; } long length = -1; try { length = Files.size(tmp); } catch (IOException e) { LOGGER.warn("IOException during digesting: " + tmp.toAbsolutePath()); return; } String targetPath = targetPathRewriter.rewrite(targetDigest); Metadata metadata = new Metadata(); try (InputStream is = TikaInputStream.get(tmp, metadata)) { emitter.emit(targetPath, is, new Metadata()); logSuccess(ccIndexRecord, targetDigest, length, targetPath); } catch (IOException | TikaException e) { LOGGER.warn("problem writing id={}", id, e); } } finally { try { Files.delete(tmp); } catch (IOException e) { LOGGER.warn("can't delete " + tmp.toAbsolutePath(), e); } } } private void logSuccess(CCIndexRecord ccIndexRecord, String targetDigest, long length, String targetPath) { if (extractTruncated) { EXTRACTED_ALL_LOGGER.info("", ccIndexRecord.getUrl(), ccIndexRecord.getNormalizedMime(), ccIndexRecord.getNormalizedMimeDetected(), ccIndexRecord.getFilename(), ccIndexRecord
.getOffset(), ccIndexRecord.getLength(), ccIndexRecord.getTruncated(), targetDigest, length, targetPath);
} else { //new ObjectArray ? //url,mime_detected,warc_file,warc_offset,warc_length,sha256,length,path EXTRACTED_LOGGER.info("", ccIndexRecord.getUrl(), ccIndexRecord.getNormalizedMime(), ccIndexRecord.getNormalizedMimeDetected(), ccIndexRecord.getFilename(), ccIndexRecord.getOffset(), ccIndexRecord.getLength(), targetDigest, length, targetPath); } } private void parseWarc(String id, CCIndexRecord ccIndexRecord, byte[] warcRecordGZBytes) throws IOException { //need to leave initial inputstream open while parsing warcrecord //can't just parse record and return try (InputStream is = new GZIPInputStream(new ByteArrayInputStream(warcRecordGZBytes))) { try (WarcReader warcreader = new WarcReader(is)) { //should be a single warc per file //return the first for (WarcRecord warcRecord : warcreader) { fetchPayload(id, ccIndexRecord, warcRecord); return; } } } } private byte[] fetchWarcBytes(FetchEmitTuple t) throws TikaException, InterruptedException, IOException { ByteArrayOutputStream bos = new ByteArrayOutputStream(); FetchKey k = t.getFetchKey(); try (InputStream is = fetcher.fetch(k.getFetchKey(), k.getRangeStart(), k.getRangeEnd(), new Metadata())) { IOUtils.copy(is, bos); } return bos.toByteArray(); } }
src/main/java/org/tallison/cc/index/extractor/FileFromCCWarcExtractor.java
tballison-commoncrawl-fetcher-lite-2c106d1
[ { "filename": "src/main/java/org/tallison/cc/index/extractor/CCFileExtractorRecordProcessor.java", "retrieved_chunk": "import org.slf4j.LoggerFactory;\nimport org.tallison.cc.index.AbstractRecordProcessor;\nimport org.tallison.cc.index.CCIndexReaderCounter;\nimport org.tallison.cc.index.CCIndexRecord;\nimport org.apache.tika.exception.TikaConfigException;\nimport org.apache.tika.utils.StringUtils;\npublic class CCFileExtractorRecordProcessor extends AbstractRecordProcessor {\n private static Logger LOGGER = LoggerFactory.getLogger(CCFileExtractorRecordProcessor.class);\n private static Logger TRUNCATED_URLS_LOGGER = LoggerFactory.getLogger(\"truncated-urls\");\n private static Logger TRUNCATED_URLS_FULL_LOGGER =", "score": 0.8587356805801392 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/CCFileExtractor.java", "retrieved_chunk": "import org.apache.tika.utils.StringUtils;\n/**\n * This is a lighter class that doesn't rely on a database\n * to extract files from CC and log a list of truncated urls.\n */\npublic class CCFileExtractor {\n private static final Long INDEX_WORKER_ID = 1l;\n private static final Long INDEX_READER_ID = 2l;\n private static final Logger LOGGER = LoggerFactory.getLogger(CCFileExtractor.class);\n public static void main(String[] args) throws Exception {", "score": 0.8386580944061279 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/CCFileExtractorRecordProcessor.java", "retrieved_chunk": " String url = r.getUrl();\n TRUNCATED_URLS_LOGGER.info(\"\", url);\n //url,mime,mime_detected,warc_file,warc_offset,warc_length,truncated\n TRUNCATED_URLS_FULL_LOGGER.info(\"\", url,\n r.getNormalizedMime(), r.getNormalizedMimeDetected(), r.getFilename(),\n r.getOffset(), r.getLength(), r.getTruncated());\n return true;\n }\n }\n private void fetchBytes(CCIndexRecord r) throws InterruptedException {", "score": 0.8355271220207214 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/CCFileExtractorRecordProcessor.java", "retrieved_chunk": " LoggerFactory.getLogger(\"truncated-urls-full\");\n private final ExtractorConfig fetcherConfig;\n private final CCIndexReaderCounter counter;\n private final FileFromCCWarcExtractor fileFromCCWarcFetcher;\n private long reportEvery = 100000;\n public CCFileExtractorRecordProcessor(ExtractorConfig fetcherConfig, CCIndexReaderCounter counter)\n throws TikaConfigException, IOException {\n this.fetcherConfig = fetcherConfig;\n this.counter = counter;\n this.fileFromCCWarcFetcher = new FileFromCCWarcExtractor(fetcherConfig, counter);", "score": 0.8340771794319153 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/CCMimeCounter.java", "retrieved_chunk": " return true;\n }\n }\n private static class DetectedMimeCounter extends AbstractRecordProcessor {\n private final ExtractorConfig fetcherConfig;\n private final CCIndexReaderCounter counter;\n private final Map<String, MutableLong> totalCounts = new HashMap<>();\n private final Map<String, MutableLong> truncatedCounts = new HashMap<>();\n public DetectedMimeCounter(ExtractorConfig fetcherConfig, CCIndexReaderCounter counter) {\n this.fetcherConfig = fetcherConfig;", "score": 0.8156695365905762 } ]
java
.getOffset(), ccIndexRecord.getLength(), ccIndexRecord.getTruncated(), targetDigest, length, targetPath);
/* * Licensed to the Apache Software Foundation (ASF) under one or more * contributor license agreements. See the NOTICE file distributed with * this work for additional information regarding copyright ownership. * The ASF licenses this file to You under the Apache License, Version 2.0 * (the "License"); you may not use this file except in compliance with * the License. You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ package org.tallison.cc.index.extractor; import java.io.ByteArrayInputStream; import java.io.ByteArrayOutputStream; import java.io.IOException; import java.io.InputStream; import java.nio.file.Files; import java.nio.file.Path; import java.nio.file.StandardCopyOption; import java.util.Optional; import java.util.zip.GZIPInputStream; import org.apache.commons.codec.binary.Base32; import org.apache.commons.codec.digest.DigestUtils; import org.apache.commons.io.IOUtils; import org.netpreserve.jwarc.MediaType; import org.netpreserve.jwarc.WarcPayload; import org.netpreserve.jwarc.WarcReader; import org.netpreserve.jwarc.WarcRecord; import org.netpreserve.jwarc.WarcResponse; import org.slf4j.Logger; import org.slf4j.LoggerFactory; import org.tallison.cc.index.CCIndexReaderCounter; import org.tallison.cc.index.CCIndexRecord; import org.tallison.cc.index.io.TargetPathRewriter; import org.apache.tika.exception.TikaConfigException; import org.apache.tika.exception.TikaException; import org.apache.tika.io.TikaInputStream; import org.apache.tika.metadata.Metadata; import org.apache.tika.pipes.FetchEmitTuple; import org.apache.tika.pipes.emitter.EmitKey; import org.apache.tika.pipes.emitter.StreamEmitter; import org.apache.tika.pipes.fetcher.FetchKey; import org.apache.tika.pipes.fetcher.RangeFetcher; public class FileFromCCWarcExtractor { private static Logger LOGGER = LoggerFactory.getLogger(FileFromCCWarcExtractor.class); private static Logger EXTRACTED_LOGGER = LoggerFactory.getLogger("extracted-urls"); private static Logger EXTRACTED_ALL_LOGGER = LoggerFactory.getLogger("extracted-urls-all"); private final StreamEmitter emitter; private final TargetPathRewriter targetPathRewriter; private RangeFetcher fetcher; private final boolean extractTruncated; private Base32 base32 = new Base32(); private final CCIndexReaderCounter ccIndexReaderCounter; public FileFromCCWarcExtractor(ExtractorConfig fetcherConfig, CCIndexReaderCounter ccIndexReaderCounter) throws TikaConfigException { this.emitter = fetcherConfig.newEmitter(); this.fetcher = (RangeFetcher) fetcherConfig.newFetcher(); this.targetPathRewriter = fetcherConfig.getTargetPathRewriter(); this.extractTruncated = fetcherConfig.isExtractTruncated(); this.ccIndexReaderCounter = ccIndexReaderCounter; } public void fetchToPath(CCIndexRecord record) throws InterruptedException { LOGGER.debug("going to fetch {} {}->{}", record.getFilename(), record.getOffset(), record.getLength()); FetchEmitTuple t = new FetchEmitTuple(record.getFilename(), new FetchKey("", record.getFilename(), record.getOffset(), record.getOffset() + record.getLength() - 1), new EmitKey()); byte[] warcRecordGZBytes; try { warcRecordGZBytes = fetchWarcBytes(t); } catch (TikaException | IOException e) { LOGGER.warn("couldn't get bytes from cc's warc " + t, e); return; } String id = record.getUrl(); try { parseWarc(id, record, warcRecordGZBytes); } catch (IOException e) { LOGGER.warn("problem parsing warc file", e); } } private void fetchPayload(String id, CCIndexRecord ccIndexRecord, WarcRecord record) throws IOException { if (!((record instanceof WarcResponse) && record.contentType().base().equals(MediaType.HTTP))) { return; } Optional<WarcPayload> payload = ((WarcResponse) record).payload(); if (!payload.isPresent()) { LOGGER.debug("no payload {}", id); ccIndexReaderCounter.getEmptyPayload().incrementAndGet(); return; } if (payload.get().body().size() == 0) { LOGGER.debug("empty payload id={}", id); ccIndexReaderCounter.getEmptyPayload().incrementAndGet(); return; } Path tmp = Files.createTempFile("ccfile-fetcher-", ""); try { Files.copy(payload.get().body().stream(), tmp, StandardCopyOption.REPLACE_EXISTING); String targetDigest = null; String base32Sha1 = ""; try (InputStream is = Files.newInputStream(tmp)) { base32Sha1 = base32.encodeAsString(DigestUtils.sha1(is)); } catch (IOException e) { LOGGER.warn("IOException during digesting: " + tmp.toAbsolutePath()); return; } if (!base32Sha1.equals(ccIndexRecord.getDigest())) { LOGGER.warn("Bad digest for url={} ccindex={} sha1={}", id, ccIndexRecord.getDigest(), base32Sha1); } //TODO: make digest and encoding configurable try (InputStream is = Files.newInputStream(tmp)) { targetDigest = DigestUtils.sha256Hex(is); } catch (IOException e) { LOGGER.warn("IOException during digesting: " + tmp.toAbsolutePath()); return; } long length = -1; try { length = Files.size(tmp); } catch (IOException e) { LOGGER.warn("IOException during digesting: " + tmp.toAbsolutePath()); return; } String targetPath = targetPathRewriter.rewrite(targetDigest); Metadata metadata = new Metadata(); try (InputStream is = TikaInputStream.get(tmp, metadata)) { emitter.emit(targetPath, is, new Metadata()); logSuccess(ccIndexRecord, targetDigest, length, targetPath); } catch (IOException | TikaException e) { LOGGER.warn("problem writing id={}", id, e); } } finally { try { Files.delete(tmp); } catch (IOException e) { LOGGER.warn("can't delete " + tmp.toAbsolutePath(), e); } } } private void logSuccess(CCIndexRecord ccIndexRecord, String targetDigest, long length, String targetPath) { if (extractTruncated) { EXTRACTED_ALL_LOGGER.info("", ccIndexRecord.getUrl(), ccIndexRecord.getNormalizedMime(), ccIndexRecord.getNormalizedMimeDetected(), ccIndexRecord.getFilename(), ccIndexRecord.getOffset(), ccIndexRecord.getLength(), ccIndexRecord.getTruncated(), targetDigest, length, targetPath); } else { //new ObjectArray ? //url,mime_detected,warc_file,warc_offset,warc_length,sha256,length,path EXTRACTED_LOGGER.info("", ccIndexRecord.getUrl(), ccIndexRecord.getNormalizedMime(), ccIndexRecord.getNormalizedMimeDetected(), ccIndexRecord.getFilename(),
ccIndexRecord.getOffset(), ccIndexRecord.getLength(), targetDigest, length, targetPath);
} } private void parseWarc(String id, CCIndexRecord ccIndexRecord, byte[] warcRecordGZBytes) throws IOException { //need to leave initial inputstream open while parsing warcrecord //can't just parse record and return try (InputStream is = new GZIPInputStream(new ByteArrayInputStream(warcRecordGZBytes))) { try (WarcReader warcreader = new WarcReader(is)) { //should be a single warc per file //return the first for (WarcRecord warcRecord : warcreader) { fetchPayload(id, ccIndexRecord, warcRecord); return; } } } } private byte[] fetchWarcBytes(FetchEmitTuple t) throws TikaException, InterruptedException, IOException { ByteArrayOutputStream bos = new ByteArrayOutputStream(); FetchKey k = t.getFetchKey(); try (InputStream is = fetcher.fetch(k.getFetchKey(), k.getRangeStart(), k.getRangeEnd(), new Metadata())) { IOUtils.copy(is, bos); } return bos.toByteArray(); } }
src/main/java/org/tallison/cc/index/extractor/FileFromCCWarcExtractor.java
tballison-commoncrawl-fetcher-lite-2c106d1
[ { "filename": "src/main/java/org/tallison/cc/index/extractor/CCFileExtractorRecordProcessor.java", "retrieved_chunk": " String url = r.getUrl();\n TRUNCATED_URLS_LOGGER.info(\"\", url);\n //url,mime,mime_detected,warc_file,warc_offset,warc_length,truncated\n TRUNCATED_URLS_FULL_LOGGER.info(\"\", url,\n r.getNormalizedMime(), r.getNormalizedMimeDetected(), r.getFilename(),\n r.getOffset(), r.getLength(), r.getTruncated());\n return true;\n }\n }\n private void fetchBytes(CCIndexRecord r) throws InterruptedException {", "score": 0.8485264778137207 }, { "filename": "src/main/java/org/tallison/cc/index/CCIndexRecord.java", "retrieved_chunk": "import com.fasterxml.jackson.databind.PropertyNamingStrategies;\nimport com.fasterxml.jackson.databind.annotation.JsonNaming;\nimport org.slf4j.Logger;\nimport org.slf4j.LoggerFactory;\n@JsonNaming(PropertyNamingStrategies.KebabCaseStrategy.class)\npublic class CCIndexRecord {\n private static Pattern INT_PATTERN = Pattern.compile(\"^\\\\d+$\");\n private static Logger LOGGER = LoggerFactory.getLogger(CCIndexRecord.class);\n private static ObjectMapper OBJECT_MAPPER = new ObjectMapper();\n private String url;", "score": 0.8469132781028748 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/CCFileExtractorRecordProcessor.java", "retrieved_chunk": "import org.slf4j.LoggerFactory;\nimport org.tallison.cc.index.AbstractRecordProcessor;\nimport org.tallison.cc.index.CCIndexReaderCounter;\nimport org.tallison.cc.index.CCIndexRecord;\nimport org.apache.tika.exception.TikaConfigException;\nimport org.apache.tika.utils.StringUtils;\npublic class CCFileExtractorRecordProcessor extends AbstractRecordProcessor {\n private static Logger LOGGER = LoggerFactory.getLogger(CCFileExtractorRecordProcessor.class);\n private static Logger TRUNCATED_URLS_LOGGER = LoggerFactory.getLogger(\"truncated-urls\");\n private static Logger TRUNCATED_URLS_FULL_LOGGER =", "score": 0.8446638584136963 }, { "filename": "src/main/java/org/tallison/cc/index/CCIndexRecord.java", "retrieved_chunk": " return \"bytes=\" + offset + \"-\" + (offset + length - 1);\n }\n @Override\n public String toString() {\n return \"CCIndexRecord{\" + \"url='\" + url + '\\'' + \", mime='\" + mime + '\\'' +\n \", mimeDetected='\" + mimeDetected + '\\'' + \", status=\" + status + \", digest='\" +\n digest + '\\'' + \", length=\" + length + \", offset=\" + offset + \", filename='\" +\n filename + '\\'' + \", charset='\" + charset + '\\'' + \", languages='\" + languages +\n '\\'' + \", truncated='\" + truncated + '\\'' + \", redirect='\" + redirect + '\\'' + '}';\n }", "score": 0.8328418135643005 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/CCFileExtractor.java", "retrieved_chunk": "import org.apache.tika.utils.StringUtils;\n/**\n * This is a lighter class that doesn't rely on a database\n * to extract files from CC and log a list of truncated urls.\n */\npublic class CCFileExtractor {\n private static final Long INDEX_WORKER_ID = 1l;\n private static final Long INDEX_READER_ID = 2l;\n private static final Logger LOGGER = LoggerFactory.getLogger(CCFileExtractor.class);\n public static void main(String[] args) throws Exception {", "score": 0.8293375372886658 } ]
java
ccIndexRecord.getOffset(), ccIndexRecord.getLength(), targetDigest, length, targetPath);
/* * Licensed to the Apache Software Foundation (ASF) under one or more * contributor license agreements. See the NOTICE file distributed with * this work for additional information regarding copyright ownership. * The ASF licenses this file to You under the Apache License, Version 2.0 * (the "License"); you may not use this file except in compliance with * the License. You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ package org.tallison.cc.index.extractor; import java.io.ByteArrayInputStream; import java.io.ByteArrayOutputStream; import java.io.IOException; import java.io.InputStream; import java.nio.file.Files; import java.nio.file.Path; import java.nio.file.StandardCopyOption; import java.util.Optional; import java.util.zip.GZIPInputStream; import org.apache.commons.codec.binary.Base32; import org.apache.commons.codec.digest.DigestUtils; import org.apache.commons.io.IOUtils; import org.netpreserve.jwarc.MediaType; import org.netpreserve.jwarc.WarcPayload; import org.netpreserve.jwarc.WarcReader; import org.netpreserve.jwarc.WarcRecord; import org.netpreserve.jwarc.WarcResponse; import org.slf4j.Logger; import org.slf4j.LoggerFactory; import org.tallison.cc.index.CCIndexReaderCounter; import org.tallison.cc.index.CCIndexRecord; import org.tallison.cc.index.io.TargetPathRewriter; import org.apache.tika.exception.TikaConfigException; import org.apache.tika.exception.TikaException; import org.apache.tika.io.TikaInputStream; import org.apache.tika.metadata.Metadata; import org.apache.tika.pipes.FetchEmitTuple; import org.apache.tika.pipes.emitter.EmitKey; import org.apache.tika.pipes.emitter.StreamEmitter; import org.apache.tika.pipes.fetcher.FetchKey; import org.apache.tika.pipes.fetcher.RangeFetcher; public class FileFromCCWarcExtractor { private static Logger LOGGER = LoggerFactory.getLogger(FileFromCCWarcExtractor.class); private static Logger EXTRACTED_LOGGER = LoggerFactory.getLogger("extracted-urls"); private static Logger EXTRACTED_ALL_LOGGER = LoggerFactory.getLogger("extracted-urls-all"); private final StreamEmitter emitter; private final TargetPathRewriter targetPathRewriter; private RangeFetcher fetcher; private final boolean extractTruncated; private Base32 base32 = new Base32(); private final CCIndexReaderCounter ccIndexReaderCounter; public FileFromCCWarcExtractor(ExtractorConfig fetcherConfig, CCIndexReaderCounter ccIndexReaderCounter) throws TikaConfigException { this.emitter = fetcherConfig.newEmitter(); this.fetcher = (RangeFetcher) fetcherConfig.newFetcher(); this.targetPathRewriter = fetcherConfig.getTargetPathRewriter(); this.extractTruncated = fetcherConfig.isExtractTruncated(); this.ccIndexReaderCounter = ccIndexReaderCounter; } public void fetchToPath(CCIndexRecord record) throws InterruptedException { LOGGER.debug("going to fetch {} {}->{}", record.getFilename(), record.getOffset(), record.getLength()); FetchEmitTuple t = new FetchEmitTuple(record.getFilename(), new FetchKey("", record.getFilename(), record.getOffset(), record.getOffset() + record.getLength() - 1), new EmitKey()); byte[] warcRecordGZBytes; try { warcRecordGZBytes = fetchWarcBytes(t); } catch (TikaException | IOException e) { LOGGER.warn("couldn't get bytes from cc's warc " + t, e); return; } String id = record.getUrl(); try { parseWarc(id, record, warcRecordGZBytes); } catch (IOException e) { LOGGER.warn("problem parsing warc file", e); } } private void fetchPayload(String id, CCIndexRecord ccIndexRecord, WarcRecord record) throws IOException { if (!((record instanceof WarcResponse) && record.contentType().base().equals(MediaType.HTTP))) { return; } Optional<WarcPayload> payload = ((WarcResponse) record).payload(); if (!payload.isPresent()) { LOGGER.debug("no payload {}", id); ccIndexReaderCounter.getEmptyPayload().incrementAndGet(); return; } if (payload.get().body().size() == 0) { LOGGER.debug("empty payload id={}", id); ccIndexReaderCounter.getEmptyPayload().incrementAndGet(); return; } Path tmp = Files.createTempFile("ccfile-fetcher-", ""); try { Files.copy(payload.get().body().stream(), tmp, StandardCopyOption.REPLACE_EXISTING); String targetDigest = null; String base32Sha1 = ""; try (InputStream is = Files.newInputStream(tmp)) { base32Sha1 = base32.encodeAsString(DigestUtils.sha1(is)); } catch (IOException e) { LOGGER.warn("IOException during digesting: " + tmp.toAbsolutePath()); return; } if (!base32Sha1.equals(ccIndexRecord.getDigest())) { LOGGER.warn("Bad digest for url={} ccindex={} sha1={}", id, ccIndexRecord.getDigest(), base32Sha1); } //TODO: make digest and encoding configurable try (InputStream is = Files.newInputStream(tmp)) { targetDigest = DigestUtils.sha256Hex(is); } catch (IOException e) { LOGGER.warn("IOException during digesting: " + tmp.toAbsolutePath()); return; } long length = -1; try { length = Files.size(tmp); } catch (IOException e) { LOGGER.warn("IOException during digesting: " + tmp.toAbsolutePath()); return; } String
targetPath = targetPathRewriter.rewrite(targetDigest);
Metadata metadata = new Metadata(); try (InputStream is = TikaInputStream.get(tmp, metadata)) { emitter.emit(targetPath, is, new Metadata()); logSuccess(ccIndexRecord, targetDigest, length, targetPath); } catch (IOException | TikaException e) { LOGGER.warn("problem writing id={}", id, e); } } finally { try { Files.delete(tmp); } catch (IOException e) { LOGGER.warn("can't delete " + tmp.toAbsolutePath(), e); } } } private void logSuccess(CCIndexRecord ccIndexRecord, String targetDigest, long length, String targetPath) { if (extractTruncated) { EXTRACTED_ALL_LOGGER.info("", ccIndexRecord.getUrl(), ccIndexRecord.getNormalizedMime(), ccIndexRecord.getNormalizedMimeDetected(), ccIndexRecord.getFilename(), ccIndexRecord.getOffset(), ccIndexRecord.getLength(), ccIndexRecord.getTruncated(), targetDigest, length, targetPath); } else { //new ObjectArray ? //url,mime_detected,warc_file,warc_offset,warc_length,sha256,length,path EXTRACTED_LOGGER.info("", ccIndexRecord.getUrl(), ccIndexRecord.getNormalizedMime(), ccIndexRecord.getNormalizedMimeDetected(), ccIndexRecord.getFilename(), ccIndexRecord.getOffset(), ccIndexRecord.getLength(), targetDigest, length, targetPath); } } private void parseWarc(String id, CCIndexRecord ccIndexRecord, byte[] warcRecordGZBytes) throws IOException { //need to leave initial inputstream open while parsing warcrecord //can't just parse record and return try (InputStream is = new GZIPInputStream(new ByteArrayInputStream(warcRecordGZBytes))) { try (WarcReader warcreader = new WarcReader(is)) { //should be a single warc per file //return the first for (WarcRecord warcRecord : warcreader) { fetchPayload(id, ccIndexRecord, warcRecord); return; } } } } private byte[] fetchWarcBytes(FetchEmitTuple t) throws TikaException, InterruptedException, IOException { ByteArrayOutputStream bos = new ByteArrayOutputStream(); FetchKey k = t.getFetchKey(); try (InputStream is = fetcher.fetch(k.getFetchKey(), k.getRangeStart(), k.getRangeEnd(), new Metadata())) { IOUtils.copy(is, bos); } return bos.toByteArray(); } }
src/main/java/org/tallison/cc/index/extractor/FileFromCCWarcExtractor.java
tballison-commoncrawl-fetcher-lite-2c106d1
[ { "filename": "src/main/java/org/tallison/cc/index/CCIndexRecord.java", "retrieved_chunk": " }\n public String getDigest() {\n return digest;\n }\n public void setDigest(String digest) {\n this.digest = digest;\n }\n public Long getLength() {\n return length;\n }", "score": 0.7882855534553528 }, { "filename": "src/test/java/org/tallison/cc/index/io/TargetPathRewriterTest.java", "retrieved_chunk": " targetPathRewriter = new TargetPathRewriter(pat);\n assertEquals(\"ab/cd/e/abcdefgh\", targetPathRewriter.rewrite(txt));\n pat = \"xx/xx//xx\";\n targetPathRewriter = new TargetPathRewriter(pat);\n assertEquals(\"ab/cd//abcdefgh\", targetPathRewriter.rewrite(txt));\n }\n}", "score": 0.785385012626648 }, { "filename": "src/main/java/org/tallison/cc/index/io/TargetPathRewriter.java", "retrieved_chunk": " List<Integer> offsets = new ArrayList<>();\n public TargetPathRewriter(String targetPathPattern) {\n if (targetPathPattern.startsWith(\"/\")) {\n throw new IllegalArgumentException(\"targetPathRewriter cannot start with '/'\");\n }\n if (targetPathPattern.endsWith(\"/\")) {\n throw new IllegalArgumentException(\"targetPathRewriter cannot end with '/'\");\n }\n int i = targetPathPattern.indexOf('/');\n int hits = 0;", "score": 0.780478835105896 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/CCMimeCounter.java", "retrieved_chunk": "import java.io.File;\nimport java.io.IOException;\nimport java.io.InputStream;\nimport java.io.InputStreamReader;\nimport java.nio.charset.StandardCharsets;\nimport java.nio.file.Files;\nimport java.nio.file.Paths;\nimport java.util.ArrayList;\nimport java.util.Collections;\nimport java.util.HashMap;", "score": 0.7773416638374329 }, { "filename": "src/main/java/org/tallison/cc/index/IndexIterator.java", "retrieved_chunk": "import java.io.InputStreamReader;\nimport java.nio.charset.StandardCharsets;\nimport java.nio.file.Files;\nimport java.nio.file.Path;\nimport java.util.ArrayList;\nimport java.util.List;\nimport java.util.Map;\nimport java.util.concurrent.TimeoutException;\nimport java.util.regex.Matcher;\nimport java.util.regex.Pattern;", "score": 0.7770801186561584 } ]
java
targetPath = targetPathRewriter.rewrite(targetDigest);
/* * Licensed to the Apache Software Foundation (ASF) under one or more * contributor license agreements. See the NOTICE file distributed with * this work for additional information regarding copyright ownership. * The ASF licenses this file to You under the Apache License, Version 2.0 * (the "License"); you may not use this file except in compliance with * the License. You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ package org.tallison.cc.index.extractor; import java.io.ByteArrayInputStream; import java.io.ByteArrayOutputStream; import java.io.IOException; import java.io.InputStream; import java.nio.file.Files; import java.nio.file.Path; import java.nio.file.StandardCopyOption; import java.util.Optional; import java.util.zip.GZIPInputStream; import org.apache.commons.codec.binary.Base32; import org.apache.commons.codec.digest.DigestUtils; import org.apache.commons.io.IOUtils; import org.netpreserve.jwarc.MediaType; import org.netpreserve.jwarc.WarcPayload; import org.netpreserve.jwarc.WarcReader; import org.netpreserve.jwarc.WarcRecord; import org.netpreserve.jwarc.WarcResponse; import org.slf4j.Logger; import org.slf4j.LoggerFactory; import org.tallison.cc.index.CCIndexReaderCounter; import org.tallison.cc.index.CCIndexRecord; import org.tallison.cc.index.io.TargetPathRewriter; import org.apache.tika.exception.TikaConfigException; import org.apache.tika.exception.TikaException; import org.apache.tika.io.TikaInputStream; import org.apache.tika.metadata.Metadata; import org.apache.tika.pipes.FetchEmitTuple; import org.apache.tika.pipes.emitter.EmitKey; import org.apache.tika.pipes.emitter.StreamEmitter; import org.apache.tika.pipes.fetcher.FetchKey; import org.apache.tika.pipes.fetcher.RangeFetcher; public class FileFromCCWarcExtractor { private static Logger LOGGER = LoggerFactory.getLogger(FileFromCCWarcExtractor.class); private static Logger EXTRACTED_LOGGER = LoggerFactory.getLogger("extracted-urls"); private static Logger EXTRACTED_ALL_LOGGER = LoggerFactory.getLogger("extracted-urls-all"); private final StreamEmitter emitter; private final TargetPathRewriter targetPathRewriter; private RangeFetcher fetcher; private final boolean extractTruncated; private Base32 base32 = new Base32(); private final CCIndexReaderCounter ccIndexReaderCounter; public FileFromCCWarcExtractor(ExtractorConfig fetcherConfig, CCIndexReaderCounter ccIndexReaderCounter) throws TikaConfigException { this.emitter = fetcherConfig.newEmitter(); this.fetcher = (
RangeFetcher) fetcherConfig.newFetcher();
this.targetPathRewriter = fetcherConfig.getTargetPathRewriter(); this.extractTruncated = fetcherConfig.isExtractTruncated(); this.ccIndexReaderCounter = ccIndexReaderCounter; } public void fetchToPath(CCIndexRecord record) throws InterruptedException { LOGGER.debug("going to fetch {} {}->{}", record.getFilename(), record.getOffset(), record.getLength()); FetchEmitTuple t = new FetchEmitTuple(record.getFilename(), new FetchKey("", record.getFilename(), record.getOffset(), record.getOffset() + record.getLength() - 1), new EmitKey()); byte[] warcRecordGZBytes; try { warcRecordGZBytes = fetchWarcBytes(t); } catch (TikaException | IOException e) { LOGGER.warn("couldn't get bytes from cc's warc " + t, e); return; } String id = record.getUrl(); try { parseWarc(id, record, warcRecordGZBytes); } catch (IOException e) { LOGGER.warn("problem parsing warc file", e); } } private void fetchPayload(String id, CCIndexRecord ccIndexRecord, WarcRecord record) throws IOException { if (!((record instanceof WarcResponse) && record.contentType().base().equals(MediaType.HTTP))) { return; } Optional<WarcPayload> payload = ((WarcResponse) record).payload(); if (!payload.isPresent()) { LOGGER.debug("no payload {}", id); ccIndexReaderCounter.getEmptyPayload().incrementAndGet(); return; } if (payload.get().body().size() == 0) { LOGGER.debug("empty payload id={}", id); ccIndexReaderCounter.getEmptyPayload().incrementAndGet(); return; } Path tmp = Files.createTempFile("ccfile-fetcher-", ""); try { Files.copy(payload.get().body().stream(), tmp, StandardCopyOption.REPLACE_EXISTING); String targetDigest = null; String base32Sha1 = ""; try (InputStream is = Files.newInputStream(tmp)) { base32Sha1 = base32.encodeAsString(DigestUtils.sha1(is)); } catch (IOException e) { LOGGER.warn("IOException during digesting: " + tmp.toAbsolutePath()); return; } if (!base32Sha1.equals(ccIndexRecord.getDigest())) { LOGGER.warn("Bad digest for url={} ccindex={} sha1={}", id, ccIndexRecord.getDigest(), base32Sha1); } //TODO: make digest and encoding configurable try (InputStream is = Files.newInputStream(tmp)) { targetDigest = DigestUtils.sha256Hex(is); } catch (IOException e) { LOGGER.warn("IOException during digesting: " + tmp.toAbsolutePath()); return; } long length = -1; try { length = Files.size(tmp); } catch (IOException e) { LOGGER.warn("IOException during digesting: " + tmp.toAbsolutePath()); return; } String targetPath = targetPathRewriter.rewrite(targetDigest); Metadata metadata = new Metadata(); try (InputStream is = TikaInputStream.get(tmp, metadata)) { emitter.emit(targetPath, is, new Metadata()); logSuccess(ccIndexRecord, targetDigest, length, targetPath); } catch (IOException | TikaException e) { LOGGER.warn("problem writing id={}", id, e); } } finally { try { Files.delete(tmp); } catch (IOException e) { LOGGER.warn("can't delete " + tmp.toAbsolutePath(), e); } } } private void logSuccess(CCIndexRecord ccIndexRecord, String targetDigest, long length, String targetPath) { if (extractTruncated) { EXTRACTED_ALL_LOGGER.info("", ccIndexRecord.getUrl(), ccIndexRecord.getNormalizedMime(), ccIndexRecord.getNormalizedMimeDetected(), ccIndexRecord.getFilename(), ccIndexRecord.getOffset(), ccIndexRecord.getLength(), ccIndexRecord.getTruncated(), targetDigest, length, targetPath); } else { //new ObjectArray ? //url,mime_detected,warc_file,warc_offset,warc_length,sha256,length,path EXTRACTED_LOGGER.info("", ccIndexRecord.getUrl(), ccIndexRecord.getNormalizedMime(), ccIndexRecord.getNormalizedMimeDetected(), ccIndexRecord.getFilename(), ccIndexRecord.getOffset(), ccIndexRecord.getLength(), targetDigest, length, targetPath); } } private void parseWarc(String id, CCIndexRecord ccIndexRecord, byte[] warcRecordGZBytes) throws IOException { //need to leave initial inputstream open while parsing warcrecord //can't just parse record and return try (InputStream is = new GZIPInputStream(new ByteArrayInputStream(warcRecordGZBytes))) { try (WarcReader warcreader = new WarcReader(is)) { //should be a single warc per file //return the first for (WarcRecord warcRecord : warcreader) { fetchPayload(id, ccIndexRecord, warcRecord); return; } } } } private byte[] fetchWarcBytes(FetchEmitTuple t) throws TikaException, InterruptedException, IOException { ByteArrayOutputStream bos = new ByteArrayOutputStream(); FetchKey k = t.getFetchKey(); try (InputStream is = fetcher.fetch(k.getFetchKey(), k.getRangeStart(), k.getRangeEnd(), new Metadata())) { IOUtils.copy(is, bos); } return bos.toByteArray(); } }
src/main/java/org/tallison/cc/index/extractor/FileFromCCWarcExtractor.java
tballison-commoncrawl-fetcher-lite-2c106d1
[ { "filename": "src/main/java/org/tallison/cc/index/extractor/CCFileExtractorRecordProcessor.java", "retrieved_chunk": " LoggerFactory.getLogger(\"truncated-urls-full\");\n private final ExtractorConfig fetcherConfig;\n private final CCIndexReaderCounter counter;\n private final FileFromCCWarcExtractor fileFromCCWarcFetcher;\n private long reportEvery = 100000;\n public CCFileExtractorRecordProcessor(ExtractorConfig fetcherConfig, CCIndexReaderCounter counter)\n throws TikaConfigException, IOException {\n this.fetcherConfig = fetcherConfig;\n this.counter = counter;\n this.fileFromCCWarcFetcher = new FileFromCCWarcExtractor(fetcherConfig, counter);", "score": 0.9185293912887573 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/ExtractorConfig.java", "retrieved_chunk": "import com.fasterxml.jackson.annotation.JsonProperty;\nimport org.tallison.cc.index.IndexIterator;\nimport org.tallison.cc.index.io.BackoffHttpFetcher;\nimport org.tallison.cc.index.io.TargetPathRewriter;\nimport org.tallison.cc.index.selector.RecordSelector;\nimport org.apache.tika.config.Initializable;\nimport org.apache.tika.exception.TikaConfigException;\nimport org.apache.tika.pipes.emitter.StreamEmitter;\nimport org.apache.tika.pipes.emitter.fs.FileSystemEmitter;\nimport org.apache.tika.pipes.emitter.s3.S3Emitter;", "score": 0.8590492010116577 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/CCFileExtractor.java", "retrieved_chunk": " ExtractorConfig fetcherConfig =\n new ObjectMapper().readValue(new File(args[0]), ExtractorConfig.class);\n execute(fetcherConfig);\n }\n private static void execute(ExtractorConfig fetcherConfig) throws TikaException {\n ArrayBlockingQueue<FetchEmitTuple> indexPathsList = new ArrayBlockingQueue<>(1000);\n //IndexPathsReader reads a file containing a list of cc-index.paths files\n //and writes the literal gz files (cc-index/collections/CC-MAIN-2023-06/indexes/cdx-00000.gz)\n //to indexPathsList\n //IndexWorker reads a single index.gz file at a time and processes each record", "score": 0.853892982006073 }, { "filename": "src/main/java/org/tallison/cc/index/IndexIterator.java", "retrieved_chunk": "import java.util.zip.GZIPInputStream;\nimport com.fasterxml.jackson.annotation.JsonCreator;\nimport com.fasterxml.jackson.annotation.JsonProperty;\nimport org.tallison.cc.index.extractor.ExtractorConfig;\nimport org.tallison.cc.index.io.BackoffHttpFetcher;\nimport org.apache.tika.config.Initializable;\nimport org.apache.tika.config.Param;\nimport org.apache.tika.exception.TikaConfigException;\nimport org.apache.tika.exception.TikaException;\nimport org.apache.tika.metadata.Metadata;", "score": 0.8513529300689697 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/CCFileExtractor.java", "retrieved_chunk": "import org.tallison.cc.index.CCIndexReaderCounter;\nimport org.tallison.cc.index.IndexIterator;\nimport org.apache.tika.exception.TikaConfigException;\nimport org.apache.tika.exception.TikaException;\nimport org.apache.tika.io.TikaInputStream;\nimport org.apache.tika.metadata.Metadata;\nimport org.apache.tika.pipes.FetchEmitTuple;\nimport org.apache.tika.pipes.fetcher.Fetcher;\nimport org.apache.tika.pipes.pipesiterator.CallablePipesIterator;\nimport org.apache.tika.pipes.pipesiterator.PipesIterator;", "score": 0.850680947303772 } ]
java
RangeFetcher) fetcherConfig.newFetcher();
/* * Licensed to the Apache Software Foundation (ASF) under one or more * contributor license agreements. See the NOTICE file distributed with * this work for additional information regarding copyright ownership. * The ASF licenses this file to You under the Apache License, Version 2.0 * (the "License"); you may not use this file except in compliance with * the License. You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ package org.tallison.cc.index.extractor; import java.io.ByteArrayInputStream; import java.io.ByteArrayOutputStream; import java.io.IOException; import java.io.InputStream; import java.nio.file.Files; import java.nio.file.Path; import java.nio.file.StandardCopyOption; import java.util.Optional; import java.util.zip.GZIPInputStream; import org.apache.commons.codec.binary.Base32; import org.apache.commons.codec.digest.DigestUtils; import org.apache.commons.io.IOUtils; import org.netpreserve.jwarc.MediaType; import org.netpreserve.jwarc.WarcPayload; import org.netpreserve.jwarc.WarcReader; import org.netpreserve.jwarc.WarcRecord; import org.netpreserve.jwarc.WarcResponse; import org.slf4j.Logger; import org.slf4j.LoggerFactory; import org.tallison.cc.index.CCIndexReaderCounter; import org.tallison.cc.index.CCIndexRecord; import org.tallison.cc.index.io.TargetPathRewriter; import org.apache.tika.exception.TikaConfigException; import org.apache.tika.exception.TikaException; import org.apache.tika.io.TikaInputStream; import org.apache.tika.metadata.Metadata; import org.apache.tika.pipes.FetchEmitTuple; import org.apache.tika.pipes.emitter.EmitKey; import org.apache.tika.pipes.emitter.StreamEmitter; import org.apache.tika.pipes.fetcher.FetchKey; import org.apache.tika.pipes.fetcher.RangeFetcher; public class FileFromCCWarcExtractor { private static Logger LOGGER = LoggerFactory.getLogger(FileFromCCWarcExtractor.class); private static Logger EXTRACTED_LOGGER = LoggerFactory.getLogger("extracted-urls"); private static Logger EXTRACTED_ALL_LOGGER = LoggerFactory.getLogger("extracted-urls-all"); private final StreamEmitter emitter; private final TargetPathRewriter targetPathRewriter; private RangeFetcher fetcher; private final boolean extractTruncated; private Base32 base32 = new Base32(); private final CCIndexReaderCounter ccIndexReaderCounter; public FileFromCCWarcExtractor(ExtractorConfig fetcherConfig, CCIndexReaderCounter ccIndexReaderCounter) throws TikaConfigException { this.emitter = fetcherConfig.newEmitter(); this.fetcher = (RangeFetcher) fetcherConfig.newFetcher(); this.targetPathRewriter = fetcherConfig.getTargetPathRewriter(); this.extractTruncated = fetcherConfig.isExtractTruncated(); this.ccIndexReaderCounter = ccIndexReaderCounter; } public void fetchToPath(CCIndexRecord record) throws InterruptedException { LOGGER.debug("going to fetch {} {}->{}", record.getFilename(), record.getOffset(), record.getLength()); FetchEmitTuple t = new FetchEmitTuple(record.getFilename(), new FetchKey("", record.getFilename(), record.getOffset(), record.getOffset() + record.getLength() - 1), new EmitKey()); byte[] warcRecordGZBytes; try { warcRecordGZBytes = fetchWarcBytes(t); } catch (TikaException | IOException e) { LOGGER.warn("couldn't get bytes from cc's warc " + t, e); return; } String id = record.getUrl(); try { parseWarc(id, record, warcRecordGZBytes); } catch (IOException e) { LOGGER.warn("problem parsing warc file", e); } } private void fetchPayload(String id, CCIndexRecord ccIndexRecord, WarcRecord record) throws IOException { if (!((record instanceof WarcResponse) && record.contentType().base().equals(MediaType.HTTP))) { return; } Optional<WarcPayload> payload = ((WarcResponse) record).payload(); if (!payload.isPresent()) { LOGGER.debug("no payload {}", id); ccIndexReaderCounter.getEmptyPayload().incrementAndGet(); return; } if (payload.get().body().size() == 0) { LOGGER.debug("empty payload id={}", id); ccIndexReaderCounter.getEmptyPayload().incrementAndGet(); return; } Path tmp = Files.createTempFile("ccfile-fetcher-", ""); try { Files.copy(payload.get().body().stream(), tmp, StandardCopyOption.REPLACE_EXISTING); String targetDigest = null; String base32Sha1 = ""; try (InputStream is = Files.newInputStream(tmp)) { base32Sha1 = base32.encodeAsString(DigestUtils.sha1(is)); } catch (IOException e) { LOGGER.warn("IOException during digesting: " + tmp.toAbsolutePath()); return; } if (!base32Sha1.equals(ccIndexRecord.getDigest())) { LOGGER.warn("Bad digest for url={} ccindex={} sha1={}", id, ccIndexRecord.getDigest(), base32Sha1); } //TODO: make digest and encoding configurable try (InputStream is = Files.newInputStream(tmp)) { targetDigest = DigestUtils.sha256Hex(is); } catch (IOException e) { LOGGER.warn("IOException during digesting: " + tmp.toAbsolutePath()); return; } long length = -1; try { length = Files.size(tmp); } catch (IOException e) { LOGGER.warn("IOException during digesting: " + tmp.toAbsolutePath()); return; } String targetPath = targetPathRewriter.rewrite(targetDigest); Metadata metadata = new Metadata(); try (InputStream is = TikaInputStream.get(tmp, metadata)) { emitter.emit(targetPath, is, new Metadata()); logSuccess(ccIndexRecord, targetDigest, length, targetPath); } catch (IOException | TikaException e) { LOGGER.warn("problem writing id={}", id, e); } } finally { try { Files.delete(tmp); } catch (IOException e) { LOGGER.warn("can't delete " + tmp.toAbsolutePath(), e); } } } private void logSuccess(CCIndexRecord ccIndexRecord, String targetDigest, long length, String targetPath) { if (extractTruncated) { EXTRACTED_ALL_LOGGER.info("", ccIndexRecord.getUrl(), ccIndexRecord.getNormalizedMime(), ccIndexRecord.getNormalizedMimeDetected(), ccIndexRecord.getFilename(), ccIndexRecord.getOffset(), ccIndexRecord.getLength(), ccIndexRecord.getTruncated(), targetDigest, length, targetPath); } else { //new ObjectArray ? //url,mime_detected,warc_file,warc_offset,warc_length,sha256,length,path EXTRACTED_LOGGER.info("", ccIndexRecord.getUrl(), ccIndexRecord.getNormalizedMime(), ccIndexRecord.getNormalizedMimeDetected(),
ccIndexRecord.getFilename(), ccIndexRecord.getOffset(), ccIndexRecord.getLength(), targetDigest, length, targetPath);
} } private void parseWarc(String id, CCIndexRecord ccIndexRecord, byte[] warcRecordGZBytes) throws IOException { //need to leave initial inputstream open while parsing warcrecord //can't just parse record and return try (InputStream is = new GZIPInputStream(new ByteArrayInputStream(warcRecordGZBytes))) { try (WarcReader warcreader = new WarcReader(is)) { //should be a single warc per file //return the first for (WarcRecord warcRecord : warcreader) { fetchPayload(id, ccIndexRecord, warcRecord); return; } } } } private byte[] fetchWarcBytes(FetchEmitTuple t) throws TikaException, InterruptedException, IOException { ByteArrayOutputStream bos = new ByteArrayOutputStream(); FetchKey k = t.getFetchKey(); try (InputStream is = fetcher.fetch(k.getFetchKey(), k.getRangeStart(), k.getRangeEnd(), new Metadata())) { IOUtils.copy(is, bos); } return bos.toByteArray(); } }
src/main/java/org/tallison/cc/index/extractor/FileFromCCWarcExtractor.java
tballison-commoncrawl-fetcher-lite-2c106d1
[ { "filename": "src/main/java/org/tallison/cc/index/extractor/CCFileExtractorRecordProcessor.java", "retrieved_chunk": " String url = r.getUrl();\n TRUNCATED_URLS_LOGGER.info(\"\", url);\n //url,mime,mime_detected,warc_file,warc_offset,warc_length,truncated\n TRUNCATED_URLS_FULL_LOGGER.info(\"\", url,\n r.getNormalizedMime(), r.getNormalizedMimeDetected(), r.getFilename(),\n r.getOffset(), r.getLength(), r.getTruncated());\n return true;\n }\n }\n private void fetchBytes(CCIndexRecord r) throws InterruptedException {", "score": 0.8481252193450928 }, { "filename": "src/main/java/org/tallison/cc/index/CCIndexRecord.java", "retrieved_chunk": "import com.fasterxml.jackson.databind.PropertyNamingStrategies;\nimport com.fasterxml.jackson.databind.annotation.JsonNaming;\nimport org.slf4j.Logger;\nimport org.slf4j.LoggerFactory;\n@JsonNaming(PropertyNamingStrategies.KebabCaseStrategy.class)\npublic class CCIndexRecord {\n private static Pattern INT_PATTERN = Pattern.compile(\"^\\\\d+$\");\n private static Logger LOGGER = LoggerFactory.getLogger(CCIndexRecord.class);\n private static ObjectMapper OBJECT_MAPPER = new ObjectMapper();\n private String url;", "score": 0.8461720943450928 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/CCFileExtractorRecordProcessor.java", "retrieved_chunk": "import org.slf4j.LoggerFactory;\nimport org.tallison.cc.index.AbstractRecordProcessor;\nimport org.tallison.cc.index.CCIndexReaderCounter;\nimport org.tallison.cc.index.CCIndexRecord;\nimport org.apache.tika.exception.TikaConfigException;\nimport org.apache.tika.utils.StringUtils;\npublic class CCFileExtractorRecordProcessor extends AbstractRecordProcessor {\n private static Logger LOGGER = LoggerFactory.getLogger(CCFileExtractorRecordProcessor.class);\n private static Logger TRUNCATED_URLS_LOGGER = LoggerFactory.getLogger(\"truncated-urls\");\n private static Logger TRUNCATED_URLS_FULL_LOGGER =", "score": 0.8441181182861328 }, { "filename": "src/main/java/org/tallison/cc/index/CCIndexRecord.java", "retrieved_chunk": " return \"bytes=\" + offset + \"-\" + (offset + length - 1);\n }\n @Override\n public String toString() {\n return \"CCIndexRecord{\" + \"url='\" + url + '\\'' + \", mime='\" + mime + '\\'' +\n \", mimeDetected='\" + mimeDetected + '\\'' + \", status=\" + status + \", digest='\" +\n digest + '\\'' + \", length=\" + length + \", offset=\" + offset + \", filename='\" +\n filename + '\\'' + \", charset='\" + charset + '\\'' + \", languages='\" + languages +\n '\\'' + \", truncated='\" + truncated + '\\'' + \", redirect='\" + redirect + '\\'' + '}';\n }", "score": 0.8325063586235046 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/CCFileExtractor.java", "retrieved_chunk": "import org.apache.tika.utils.StringUtils;\n/**\n * This is a lighter class that doesn't rely on a database\n * to extract files from CC and log a list of truncated urls.\n */\npublic class CCFileExtractor {\n private static final Long INDEX_WORKER_ID = 1l;\n private static final Long INDEX_READER_ID = 2l;\n private static final Logger LOGGER = LoggerFactory.getLogger(CCFileExtractor.class);\n public static void main(String[] args) throws Exception {", "score": 0.8292436599731445 } ]
java
ccIndexRecord.getFilename(), ccIndexRecord.getOffset(), ccIndexRecord.getLength(), targetDigest, length, targetPath);
/* * Licensed to the Apache Software Foundation (ASF) under one or more * contributor license agreements. See the NOTICE file distributed with * this work for additional information regarding copyright ownership. * The ASF licenses this file to You under the Apache License, Version 2.0 * (the "License"); you may not use this file except in compliance with * the License. You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ package org.tallison.cc.index.extractor; import java.io.ByteArrayInputStream; import java.io.ByteArrayOutputStream; import java.io.IOException; import java.io.InputStream; import java.nio.file.Files; import java.nio.file.Path; import java.nio.file.StandardCopyOption; import java.util.Optional; import java.util.zip.GZIPInputStream; import org.apache.commons.codec.binary.Base32; import org.apache.commons.codec.digest.DigestUtils; import org.apache.commons.io.IOUtils; import org.netpreserve.jwarc.MediaType; import org.netpreserve.jwarc.WarcPayload; import org.netpreserve.jwarc.WarcReader; import org.netpreserve.jwarc.WarcRecord; import org.netpreserve.jwarc.WarcResponse; import org.slf4j.Logger; import org.slf4j.LoggerFactory; import org.tallison.cc.index.CCIndexReaderCounter; import org.tallison.cc.index.CCIndexRecord; import org.tallison.cc.index.io.TargetPathRewriter; import org.apache.tika.exception.TikaConfigException; import org.apache.tika.exception.TikaException; import org.apache.tika.io.TikaInputStream; import org.apache.tika.metadata.Metadata; import org.apache.tika.pipes.FetchEmitTuple; import org.apache.tika.pipes.emitter.EmitKey; import org.apache.tika.pipes.emitter.StreamEmitter; import org.apache.tika.pipes.fetcher.FetchKey; import org.apache.tika.pipes.fetcher.RangeFetcher; public class FileFromCCWarcExtractor { private static Logger LOGGER = LoggerFactory.getLogger(FileFromCCWarcExtractor.class); private static Logger EXTRACTED_LOGGER = LoggerFactory.getLogger("extracted-urls"); private static Logger EXTRACTED_ALL_LOGGER = LoggerFactory.getLogger("extracted-urls-all"); private final StreamEmitter emitter; private final TargetPathRewriter targetPathRewriter; private RangeFetcher fetcher; private final boolean extractTruncated; private Base32 base32 = new Base32(); private final CCIndexReaderCounter ccIndexReaderCounter; public FileFromCCWarcExtractor(ExtractorConfig fetcherConfig, CCIndexReaderCounter ccIndexReaderCounter) throws TikaConfigException { this.emitter = fetcherConfig.newEmitter(); this.fetcher = (RangeFetcher) fetcherConfig.newFetcher(); this
.targetPathRewriter = fetcherConfig.getTargetPathRewriter();
this.extractTruncated = fetcherConfig.isExtractTruncated(); this.ccIndexReaderCounter = ccIndexReaderCounter; } public void fetchToPath(CCIndexRecord record) throws InterruptedException { LOGGER.debug("going to fetch {} {}->{}", record.getFilename(), record.getOffset(), record.getLength()); FetchEmitTuple t = new FetchEmitTuple(record.getFilename(), new FetchKey("", record.getFilename(), record.getOffset(), record.getOffset() + record.getLength() - 1), new EmitKey()); byte[] warcRecordGZBytes; try { warcRecordGZBytes = fetchWarcBytes(t); } catch (TikaException | IOException e) { LOGGER.warn("couldn't get bytes from cc's warc " + t, e); return; } String id = record.getUrl(); try { parseWarc(id, record, warcRecordGZBytes); } catch (IOException e) { LOGGER.warn("problem parsing warc file", e); } } private void fetchPayload(String id, CCIndexRecord ccIndexRecord, WarcRecord record) throws IOException { if (!((record instanceof WarcResponse) && record.contentType().base().equals(MediaType.HTTP))) { return; } Optional<WarcPayload> payload = ((WarcResponse) record).payload(); if (!payload.isPresent()) { LOGGER.debug("no payload {}", id); ccIndexReaderCounter.getEmptyPayload().incrementAndGet(); return; } if (payload.get().body().size() == 0) { LOGGER.debug("empty payload id={}", id); ccIndexReaderCounter.getEmptyPayload().incrementAndGet(); return; } Path tmp = Files.createTempFile("ccfile-fetcher-", ""); try { Files.copy(payload.get().body().stream(), tmp, StandardCopyOption.REPLACE_EXISTING); String targetDigest = null; String base32Sha1 = ""; try (InputStream is = Files.newInputStream(tmp)) { base32Sha1 = base32.encodeAsString(DigestUtils.sha1(is)); } catch (IOException e) { LOGGER.warn("IOException during digesting: " + tmp.toAbsolutePath()); return; } if (!base32Sha1.equals(ccIndexRecord.getDigest())) { LOGGER.warn("Bad digest for url={} ccindex={} sha1={}", id, ccIndexRecord.getDigest(), base32Sha1); } //TODO: make digest and encoding configurable try (InputStream is = Files.newInputStream(tmp)) { targetDigest = DigestUtils.sha256Hex(is); } catch (IOException e) { LOGGER.warn("IOException during digesting: " + tmp.toAbsolutePath()); return; } long length = -1; try { length = Files.size(tmp); } catch (IOException e) { LOGGER.warn("IOException during digesting: " + tmp.toAbsolutePath()); return; } String targetPath = targetPathRewriter.rewrite(targetDigest); Metadata metadata = new Metadata(); try (InputStream is = TikaInputStream.get(tmp, metadata)) { emitter.emit(targetPath, is, new Metadata()); logSuccess(ccIndexRecord, targetDigest, length, targetPath); } catch (IOException | TikaException e) { LOGGER.warn("problem writing id={}", id, e); } } finally { try { Files.delete(tmp); } catch (IOException e) { LOGGER.warn("can't delete " + tmp.toAbsolutePath(), e); } } } private void logSuccess(CCIndexRecord ccIndexRecord, String targetDigest, long length, String targetPath) { if (extractTruncated) { EXTRACTED_ALL_LOGGER.info("", ccIndexRecord.getUrl(), ccIndexRecord.getNormalizedMime(), ccIndexRecord.getNormalizedMimeDetected(), ccIndexRecord.getFilename(), ccIndexRecord.getOffset(), ccIndexRecord.getLength(), ccIndexRecord.getTruncated(), targetDigest, length, targetPath); } else { //new ObjectArray ? //url,mime_detected,warc_file,warc_offset,warc_length,sha256,length,path EXTRACTED_LOGGER.info("", ccIndexRecord.getUrl(), ccIndexRecord.getNormalizedMime(), ccIndexRecord.getNormalizedMimeDetected(), ccIndexRecord.getFilename(), ccIndexRecord.getOffset(), ccIndexRecord.getLength(), targetDigest, length, targetPath); } } private void parseWarc(String id, CCIndexRecord ccIndexRecord, byte[] warcRecordGZBytes) throws IOException { //need to leave initial inputstream open while parsing warcrecord //can't just parse record and return try (InputStream is = new GZIPInputStream(new ByteArrayInputStream(warcRecordGZBytes))) { try (WarcReader warcreader = new WarcReader(is)) { //should be a single warc per file //return the first for (WarcRecord warcRecord : warcreader) { fetchPayload(id, ccIndexRecord, warcRecord); return; } } } } private byte[] fetchWarcBytes(FetchEmitTuple t) throws TikaException, InterruptedException, IOException { ByteArrayOutputStream bos = new ByteArrayOutputStream(); FetchKey k = t.getFetchKey(); try (InputStream is = fetcher.fetch(k.getFetchKey(), k.getRangeStart(), k.getRangeEnd(), new Metadata())) { IOUtils.copy(is, bos); } return bos.toByteArray(); } }
src/main/java/org/tallison/cc/index/extractor/FileFromCCWarcExtractor.java
tballison-commoncrawl-fetcher-lite-2c106d1
[ { "filename": "src/main/java/org/tallison/cc/index/extractor/CCFileExtractorRecordProcessor.java", "retrieved_chunk": " LoggerFactory.getLogger(\"truncated-urls-full\");\n private final ExtractorConfig fetcherConfig;\n private final CCIndexReaderCounter counter;\n private final FileFromCCWarcExtractor fileFromCCWarcFetcher;\n private long reportEvery = 100000;\n public CCFileExtractorRecordProcessor(ExtractorConfig fetcherConfig, CCIndexReaderCounter counter)\n throws TikaConfigException, IOException {\n this.fetcherConfig = fetcherConfig;\n this.counter = counter;\n this.fileFromCCWarcFetcher = new FileFromCCWarcExtractor(fetcherConfig, counter);", "score": 0.9187526702880859 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/CCFileExtractor.java", "retrieved_chunk": " ExtractorConfig fetcherConfig =\n new ObjectMapper().readValue(new File(args[0]), ExtractorConfig.class);\n execute(fetcherConfig);\n }\n private static void execute(ExtractorConfig fetcherConfig) throws TikaException {\n ArrayBlockingQueue<FetchEmitTuple> indexPathsList = new ArrayBlockingQueue<>(1000);\n //IndexPathsReader reads a file containing a list of cc-index.paths files\n //and writes the literal gz files (cc-index/collections/CC-MAIN-2023-06/indexes/cdx-00000.gz)\n //to indexPathsList\n //IndexWorker reads a single index.gz file at a time and processes each record", "score": 0.8547456860542297 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/CCFileExtractor.java", "retrieved_chunk": "import org.tallison.cc.index.CCIndexReaderCounter;\nimport org.tallison.cc.index.IndexIterator;\nimport org.apache.tika.exception.TikaConfigException;\nimport org.apache.tika.exception.TikaException;\nimport org.apache.tika.io.TikaInputStream;\nimport org.apache.tika.metadata.Metadata;\nimport org.apache.tika.pipes.FetchEmitTuple;\nimport org.apache.tika.pipes.fetcher.Fetcher;\nimport org.apache.tika.pipes.pipesiterator.CallablePipesIterator;\nimport org.apache.tika.pipes.pipesiterator.PipesIterator;", "score": 0.8532994389533997 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/CCMimeCounter.java", "retrieved_chunk": "import org.tallison.cc.index.AbstractRecordProcessor;\nimport org.tallison.cc.index.CCIndexReaderCounter;\nimport org.tallison.cc.index.CCIndexRecord;\nimport org.tallison.cc.index.IndexIterator;\nimport org.apache.tika.exception.TikaConfigException;\nimport org.apache.tika.exception.TikaException;\nimport org.apache.tika.io.TikaInputStream;\nimport org.apache.tika.metadata.Metadata;\nimport org.apache.tika.pipes.FetchEmitTuple;\nimport org.apache.tika.pipes.fetcher.Fetcher;", "score": 0.8511837720870972 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/ExtractorConfig.java", "retrieved_chunk": "import com.fasterxml.jackson.annotation.JsonProperty;\nimport org.tallison.cc.index.IndexIterator;\nimport org.tallison.cc.index.io.BackoffHttpFetcher;\nimport org.tallison.cc.index.io.TargetPathRewriter;\nimport org.tallison.cc.index.selector.RecordSelector;\nimport org.apache.tika.config.Initializable;\nimport org.apache.tika.exception.TikaConfigException;\nimport org.apache.tika.pipes.emitter.StreamEmitter;\nimport org.apache.tika.pipes.emitter.fs.FileSystemEmitter;\nimport org.apache.tika.pipes.emitter.s3.S3Emitter;", "score": 0.8501334190368652 } ]
java
.targetPathRewriter = fetcherConfig.getTargetPathRewriter();
/* * Licensed to the Apache Software Foundation (ASF) under one or more * contributor license agreements. See the NOTICE file distributed with * this work for additional information regarding copyright ownership. * The ASF licenses this file to You under the Apache License, Version 2.0 * (the "License"); you may not use this file except in compliance with * the License. You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ package org.tallison.cc.index.extractor; import java.io.ByteArrayInputStream; import java.io.ByteArrayOutputStream; import java.io.IOException; import java.io.InputStream; import java.nio.file.Files; import java.nio.file.Path; import java.nio.file.StandardCopyOption; import java.util.Optional; import java.util.zip.GZIPInputStream; import org.apache.commons.codec.binary.Base32; import org.apache.commons.codec.digest.DigestUtils; import org.apache.commons.io.IOUtils; import org.netpreserve.jwarc.MediaType; import org.netpreserve.jwarc.WarcPayload; import org.netpreserve.jwarc.WarcReader; import org.netpreserve.jwarc.WarcRecord; import org.netpreserve.jwarc.WarcResponse; import org.slf4j.Logger; import org.slf4j.LoggerFactory; import org.tallison.cc.index.CCIndexReaderCounter; import org.tallison.cc.index.CCIndexRecord; import org.tallison.cc.index.io.TargetPathRewriter; import org.apache.tika.exception.TikaConfigException; import org.apache.tika.exception.TikaException; import org.apache.tika.io.TikaInputStream; import org.apache.tika.metadata.Metadata; import org.apache.tika.pipes.FetchEmitTuple; import org.apache.tika.pipes.emitter.EmitKey; import org.apache.tika.pipes.emitter.StreamEmitter; import org.apache.tika.pipes.fetcher.FetchKey; import org.apache.tika.pipes.fetcher.RangeFetcher; public class FileFromCCWarcExtractor { private static Logger LOGGER = LoggerFactory.getLogger(FileFromCCWarcExtractor.class); private static Logger EXTRACTED_LOGGER = LoggerFactory.getLogger("extracted-urls"); private static Logger EXTRACTED_ALL_LOGGER = LoggerFactory.getLogger("extracted-urls-all"); private final StreamEmitter emitter; private final TargetPathRewriter targetPathRewriter; private RangeFetcher fetcher; private final boolean extractTruncated; private Base32 base32 = new Base32(); private final CCIndexReaderCounter ccIndexReaderCounter; public FileFromCCWarcExtractor(ExtractorConfig fetcherConfig, CCIndexReaderCounter ccIndexReaderCounter) throws TikaConfigException { this.emitter = fetcherConfig.newEmitter(); this.fetcher = (RangeFetcher) fetcherConfig.newFetcher(); this.targetPathRewriter = fetcherConfig.getTargetPathRewriter(); this.extractTruncated = fetcherConfig.isExtractTruncated(); this.ccIndexReaderCounter = ccIndexReaderCounter; } public void fetchToPath(CCIndexRecord record) throws InterruptedException { LOGGER.debug("going to fetch {} {}->{}", record.getFilename(), record.getOffset(), record.getLength()); FetchEmitTuple t = new FetchEmitTuple(record.getFilename(), new FetchKey("", record.getFilename(), record.getOffset(), record.getOffset() + record.getLength() - 1), new EmitKey()); byte[] warcRecordGZBytes; try { warcRecordGZBytes = fetchWarcBytes(t); } catch (TikaException | IOException e) { LOGGER.warn("couldn't get bytes from cc's warc " + t, e); return; } String id = record.getUrl(); try { parseWarc(id, record, warcRecordGZBytes); } catch (IOException e) { LOGGER.warn("problem parsing warc file", e); } } private void fetchPayload(String id, CCIndexRecord ccIndexRecord, WarcRecord record) throws IOException { if (!((record instanceof WarcResponse) && record.contentType().base().equals(MediaType.HTTP))) { return; } Optional<WarcPayload> payload = ((WarcResponse) record).payload(); if (!payload.isPresent()) { LOGGER.debug("no payload {}", id); ccIndexReaderCounter.getEmptyPayload().incrementAndGet(); return; } if (payload.get().body().size() == 0) { LOGGER.debug("empty payload id={}", id); ccIndexReaderCounter.getEmptyPayload().incrementAndGet(); return; } Path tmp = Files.createTempFile("ccfile-fetcher-", ""); try { Files.copy(payload.get().body().stream(), tmp, StandardCopyOption.REPLACE_EXISTING); String targetDigest = null; String base32Sha1 = ""; try (InputStream is = Files.newInputStream(tmp)) { base32Sha1 = base32.encodeAsString(DigestUtils.sha1(is)); } catch (IOException e) { LOGGER.warn("IOException during digesting: " + tmp.toAbsolutePath()); return; } if (!base32Sha1.equals(ccIndexRecord.getDigest())) { LOGGER.warn("Bad digest for url={} ccindex={} sha1={}", id, ccIndexRecord.getDigest(), base32Sha1); } //TODO: make digest and encoding configurable try (InputStream is = Files.newInputStream(tmp)) { targetDigest = DigestUtils.sha256Hex(is); } catch (IOException e) { LOGGER.warn("IOException during digesting: " + tmp.toAbsolutePath()); return; } long length = -1; try { length = Files.size(tmp); } catch (IOException e) { LOGGER.warn("IOException during digesting: " + tmp.toAbsolutePath()); return; } String targetPath = targetPathRewriter.rewrite(targetDigest); Metadata metadata = new Metadata(); try (InputStream is = TikaInputStream.get(tmp, metadata)) { emitter.emit(targetPath, is, new Metadata()); logSuccess(ccIndexRecord, targetDigest, length, targetPath); } catch (IOException | TikaException e) { LOGGER.warn("problem writing id={}", id, e); } } finally { try { Files.delete(tmp); } catch (IOException e) { LOGGER.warn("can't delete " + tmp.toAbsolutePath(), e); } } } private void logSuccess(CCIndexRecord ccIndexRecord, String targetDigest, long length, String targetPath) { if (extractTruncated) { EXTRACTED_ALL_LOGGER.info("", ccIndexRecord.getUrl(), ccIndexRecord.getNormalizedMime(), ccIndexRecord.getNormalizedMimeDetected(), ccIndexRecord.getFilename(), ccIndexRecord.getOffset(), ccIndexRecord.getLength(),
ccIndexRecord.getTruncated(), targetDigest, length, targetPath);
} else { //new ObjectArray ? //url,mime_detected,warc_file,warc_offset,warc_length,sha256,length,path EXTRACTED_LOGGER.info("", ccIndexRecord.getUrl(), ccIndexRecord.getNormalizedMime(), ccIndexRecord.getNormalizedMimeDetected(), ccIndexRecord.getFilename(), ccIndexRecord.getOffset(), ccIndexRecord.getLength(), targetDigest, length, targetPath); } } private void parseWarc(String id, CCIndexRecord ccIndexRecord, byte[] warcRecordGZBytes) throws IOException { //need to leave initial inputstream open while parsing warcrecord //can't just parse record and return try (InputStream is = new GZIPInputStream(new ByteArrayInputStream(warcRecordGZBytes))) { try (WarcReader warcreader = new WarcReader(is)) { //should be a single warc per file //return the first for (WarcRecord warcRecord : warcreader) { fetchPayload(id, ccIndexRecord, warcRecord); return; } } } } private byte[] fetchWarcBytes(FetchEmitTuple t) throws TikaException, InterruptedException, IOException { ByteArrayOutputStream bos = new ByteArrayOutputStream(); FetchKey k = t.getFetchKey(); try (InputStream is = fetcher.fetch(k.getFetchKey(), k.getRangeStart(), k.getRangeEnd(), new Metadata())) { IOUtils.copy(is, bos); } return bos.toByteArray(); } }
src/main/java/org/tallison/cc/index/extractor/FileFromCCWarcExtractor.java
tballison-commoncrawl-fetcher-lite-2c106d1
[ { "filename": "src/main/java/org/tallison/cc/index/extractor/CCFileExtractorRecordProcessor.java", "retrieved_chunk": "import org.slf4j.LoggerFactory;\nimport org.tallison.cc.index.AbstractRecordProcessor;\nimport org.tallison.cc.index.CCIndexReaderCounter;\nimport org.tallison.cc.index.CCIndexRecord;\nimport org.apache.tika.exception.TikaConfigException;\nimport org.apache.tika.utils.StringUtils;\npublic class CCFileExtractorRecordProcessor extends AbstractRecordProcessor {\n private static Logger LOGGER = LoggerFactory.getLogger(CCFileExtractorRecordProcessor.class);\n private static Logger TRUNCATED_URLS_LOGGER = LoggerFactory.getLogger(\"truncated-urls\");\n private static Logger TRUNCATED_URLS_FULL_LOGGER =", "score": 0.835833728313446 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/CCFileExtractor.java", "retrieved_chunk": "import org.apache.tika.utils.StringUtils;\n/**\n * This is a lighter class that doesn't rely on a database\n * to extract files from CC and log a list of truncated urls.\n */\npublic class CCFileExtractor {\n private static final Long INDEX_WORKER_ID = 1l;\n private static final Long INDEX_READER_ID = 2l;\n private static final Logger LOGGER = LoggerFactory.getLogger(CCFileExtractor.class);\n public static void main(String[] args) throws Exception {", "score": 0.8185681700706482 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/CCFileExtractorRecordProcessor.java", "retrieved_chunk": " String url = r.getUrl();\n TRUNCATED_URLS_LOGGER.info(\"\", url);\n //url,mime,mime_detected,warc_file,warc_offset,warc_length,truncated\n TRUNCATED_URLS_FULL_LOGGER.info(\"\", url,\n r.getNormalizedMime(), r.getNormalizedMimeDetected(), r.getFilename(),\n r.getOffset(), r.getLength(), r.getTruncated());\n return true;\n }\n }\n private void fetchBytes(CCIndexRecord r) throws InterruptedException {", "score": 0.8131157159805298 }, { "filename": "src/main/java/org/tallison/cc/index/CCIndexRecord.java", "retrieved_chunk": "import com.fasterxml.jackson.databind.PropertyNamingStrategies;\nimport com.fasterxml.jackson.databind.annotation.JsonNaming;\nimport org.slf4j.Logger;\nimport org.slf4j.LoggerFactory;\n@JsonNaming(PropertyNamingStrategies.KebabCaseStrategy.class)\npublic class CCIndexRecord {\n private static Pattern INT_PATTERN = Pattern.compile(\"^\\\\d+$\");\n private static Logger LOGGER = LoggerFactory.getLogger(CCIndexRecord.class);\n private static ObjectMapper OBJECT_MAPPER = new ObjectMapper();\n private String url;", "score": 0.8049680590629578 }, { "filename": "src/test/java/org/tallison/cc/index/selector/IndexRecordSelectorTest.java", "retrieved_chunk": " public void testIndexFile() throws Exception {\n Path p = Paths.get(\"/Users/allison/data/cc/index-work/cdx-00000.gz\");\n try (BufferedReader r = new BufferedReader(\n new InputStreamReader(new GZIPInputStream(Files.newInputStream(p)),\n StandardCharsets.UTF_8))) {\n String line = r.readLine();\n while (line != null) {\n Optional<CCIndexRecord> record = CCIndexRecord.parseRecord(line);\n if (record.isPresent()) {\n CCIndexRecord indexRecord = record.get();", "score": 0.804138720035553 } ]
java
ccIndexRecord.getTruncated(), targetDigest, length, targetPath);
/* * Copyright © 2023 Deutsche Telekom AG (opensource@telekom.de) * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ package de.telekom.phonenumbernormalizer; import de.telekom.phonenumbernormalizer.dto.DeviceContext; import de.telekom.phonenumbernormalizer.dto.DeviceContextMapper; import de.telekom.phonenumbernormalizer.numberplans.PhoneLibWrapper; import de.telekom.phonenumbernormalizer.numberplans.NumberPlan; import de.telekom.phonenumbernormalizer.numberplans.NumberPlanFactory; import lombok.RequiredArgsConstructor; import org.slf4j.Logger; import org.slf4j.LoggerFactory; import org.springframework.stereotype.Component; /** * Concrete implementation of {@link PhoneNumberNormalizer} using {@link PhoneLibWrapper} to normalize a number by mitigating some inaccuracies when it comes to number plans of optional NDC and NAC as zero. * <p> * Also supports {@link DeviceContext} to enrich a phone number during normalization if the optional NDC is missing. * </p> */ @RequiredArgsConstructor @Component public class PhoneNumberNormalizerImpl implements PhoneNumberNormalizer { private static final Logger LOGGER = LoggerFactory.getLogger(PhoneNumberNormalizerImpl.class); /** * Storage for {@link PhoneNumberNormalizer#setFallbackRegionCode(String)} */ private String fallbackRegionCode = null; @Override public void setFallbackRegionCode(String fallBackRegionCode) { if (PhoneLibWrapper.getCountryCodeForRegion(fallBackRegionCode) > 0) { this.fallbackRegionCode = fallBackRegionCode; } else { this.fallbackRegionCode = null; //invalid region code! } } /** * Fallback normalization within the number-plan of the fallback region. * @param number the original number to be normalized * @param dialableNumber the original number reduced to dialable digits * @return E164 formatted phone number or at least a dialable version of the number * * @see PhoneNumberNormalizer#setFallbackRegionCode(String) * @see PhoneNumberNormalizer#normalizePhoneNumber(String, String) */ private String fallbackNormalizationFromDeviceContextToDefaultRegionCode(String number, String dialableNumber) { if (this.fallbackRegionCode == null) { LOGGER.debug("Fallback Region was set!"); return dialableNumber; } else { return this.normalizePhoneNumber(number, this.fallbackRegionCode); } } /** * Uses wrapper of PhoneLib to identify if special rules apply for normalization.<br/> * Using device context for enriching the number make it normalizable to E164 format if NDC is optional in the used number plan, but not used in the phone number to be normalized. * @param wrapper instanced wrapper of PhoneLib * @param deviceContext information like CC, NDC and {@link de.telekom.phonenumbernormalizer.dto.DeviceContextLineType} from which the number is dialled * @return E164 formatted phone number or dialable version of it or null */ private String normalize(PhoneLibWrapper wrapper, DeviceContext deviceContext) { // international prefix has been added by PhoneLib even if it's not valid in the number plan. if (wrapper == null) { LOGGER.debug("PhoneLipWrapper was not initialized"); return null; } if (wrapper.getSemiNormalizedNumber() == null) { return wrapper.getDialableNumber(); } NumberPlan numberplan = null; if (deviceContext != null) { numberplan = NumberPlanFactory
.INSTANCE.getNumberPlan(deviceContext.getLineType(), deviceContext.getCountryCode());
} if (wrapper.isShortNumber(numberplan)) { //if it is a short number, we can't add area code nor country code, so returning the dialable. return wrapper.getDialableNumber(); } if (wrapper.hasRegionNationalAccessCode() && deviceContext != null) { //Number plan is using a NationalPrefix aka Trunc Code ... so we could add Area Code if not included in the number. return wrapper.extendNumberByDefaultAreaCodeAndCountryCode(wrapper.getNationalAccessCode(), deviceContext.getNationalDestinationCode()); } // Number plan is not using NationalPrefix aka Trunc Code ... its also not a short number, so country code can be added: return wrapper.getE164Formatted(); } @Override public String normalizePhoneNumber(String number, String regionCode) { PhoneLibWrapper wrapper = new PhoneLibWrapper(number, regionCode); if (wrapper.getSemiNormalizedNumber() == null) { return wrapper.getDialableNumber(); } if (wrapper.isShortNumber()) { //if it is a short number, we can't add area code nor country code, so returning the dialable. return wrapper.getDialableNumber(); } // international prefix is added by the lib even if it's not valid in the number plan. //checking if the input number is equal to the nationalNumber based on number plan and trunc code logic. boolean hasNoCCAndNoNAC = wrapper.hasNoCountryCodeNorNationalAccessCode(); LOGGER.debug("Number has no CC and no NAC: {}.", hasNoCCAndNoNAC); //if the number is definitely a short number or needs an area code but does not have it, we do not add the country code. return (hasNoCCAndNoNAC) ? wrapper.getDialableNumber() : wrapper.getE164Formatted(); } @Override public String normalizePhoneNumber(String number, DeviceContext deviceContext) { // checking if the number has a special format or is not valid at all. PhoneLibWrapper normalizerPhoneNumber = new PhoneLibWrapper(number, null); if (! normalizerPhoneNumber.isNormalizingTried()) { return normalizerPhoneNumber.getDialableNumber(); } DeviceContext normalizedDeviceContext = DeviceContextMapper.normalized(deviceContext); if (!normalizedDeviceContext.getCountryCode().equals(DeviceContext.UNKNOWN_VALUE)) { String regionCode = PhoneLibWrapper.getRegionCodeForCountryCode(normalizedDeviceContext.getCountryCode()); // now working again with the region code normalizerPhoneNumber = new PhoneLibWrapper(number, regionCode); if (!normalizedDeviceContext.getNationalDestinationCode().equals(DeviceContext.UNKNOWN_VALUE)) { // Number needs normalization: return normalize(normalizerPhoneNumber, deviceContext); } // Device Context with CountryCode but without AreaCode ... if (!(PhoneLibWrapper.UNKNOWN_REGIONCODE.equals(regionCode))) { return this.normalizePhoneNumber(number, regionCode); } } LOGGER.debug("Normalization based on DeviceContext did not work - falling back to normalization with fallback region."); return this.fallbackNormalizationFromDeviceContextToDefaultRegionCode(number, normalizerPhoneNumber.getDialableNumber()); } }
src/main/java/de/telekom/phonenumbernormalizer/PhoneNumberNormalizerImpl.java
telekom-phonenumber-normalizer-9a1b7ad
[ { "filename": "src/main/java/de/telekom/phonenumbernormalizer/numberplans/PhoneLibWrapper.java", "retrieved_chunk": " } else {\n if (!isSpecialFormat(dialableNumber)) {\n // Number needs normalization:\n // international prefix is added by the lib even if it's not valid in the number plan.\n this.isNormalizingTried = true;\n this.semiNormalizedNumber = PhoneLibWrapper.parseNumber(dialableNumber, regionCode);\n }\n }\n }\n }", "score": 0.8675179481506348 }, { "filename": "src/main/java/de/telekom/phonenumbernormalizer/numberplans/PhoneLibWrapper.java", "retrieved_chunk": " try {\n this.semiNormalizedNumber = phoneUtil.parse(extendedNumber, regionCode);\n // after area code has been added, we can add the country code by the lib:\n return getE164Formatted();\n } catch (NumberParseException e) {\n LOGGER.warn(\"could not parse extended number: {}\", extendedNumber);\n LOGGER.debug(\"{}\", e.getMessage());\n return dialableNumber;\n }\n } else {", "score": 0.8565723896026611 }, { "filename": "src/main/java/de/telekom/phonenumbernormalizer/numberplans/PhoneLibWrapper.java", "retrieved_chunk": " String dialableNumber;\n /**\n * The given number normalized with PhoneLib, risking we get a incorrect normalization\n *\n * @see PhoneLibWrapper#PhoneLibWrapper(String, String)\n * @see PhoneLibWrapper#isNormalizingTried()\n * @see PhoneLibWrapper#getSemiNormalizedNumber()\n */\n Phonenumber.PhoneNumber semiNormalizedNumber;\n /**", "score": 0.8562072515487671 }, { "filename": "src/main/java/de/telekom/phonenumbernormalizer/numberplans/NumberPlanFactory.java", "retrieved_chunk": " LOGGER.debug(\"no number plan for country available\");\n return null;\n }\n}", "score": 0.8550909757614136 }, { "filename": "src/main/java/de/telekom/phonenumbernormalizer/numberplans/NumberPlanFactory.java", "retrieved_chunk": " */\n public NumberPlan getNumberPlan(DeviceContextLineType numberPlanType, String countryCode) {\n if (numberPlans.containsKey(numberPlanType)) {\n LOGGER.debug(\"use number plan for type: {}\", numberPlanType);\n Map<String, NumberPlan> numberPlan = numberPlans.get(numberPlanType);\n if (numberPlan.containsKey(countryCode)) {\n LOGGER.debug(\"use number plan for country code: {}\", countryCode);\n return numberPlan.get(countryCode);\n }\n }", "score": 0.8485962152481079 } ]
java
.INSTANCE.getNumberPlan(deviceContext.getLineType(), deviceContext.getCountryCode());
/* * Licensed to the Apache Software Foundation (ASF) under one or more * contributor license agreements. See the NOTICE file distributed with * this work for additional information regarding copyright ownership. * The ASF licenses this file to You under the Apache License, Version 2.0 * (the "License"); you may not use this file except in compliance with * the License. You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ package org.tallison.cc.index.extractor; import java.io.BufferedInputStream; import java.io.BufferedReader; import java.io.BufferedWriter; import java.io.File; import java.io.IOException; import java.io.InputStream; import java.io.InputStreamReader; import java.nio.charset.StandardCharsets; import java.nio.file.Files; import java.nio.file.Paths; import java.util.ArrayList; import java.util.Collections; import java.util.HashMap; import java.util.List; import java.util.Locale; import java.util.Map; import java.util.Optional; import java.util.concurrent.ArrayBlockingQueue; import java.util.concurrent.Callable; import java.util.concurrent.ExecutionException; import java.util.concurrent.ExecutorCompletionService; import java.util.concurrent.ExecutorService; import java.util.concurrent.Executors; import java.util.concurrent.Future; import java.util.concurrent.TimeUnit; import java.util.concurrent.TimeoutException; import java.util.zip.GZIPInputStream; import com.fasterxml.jackson.databind.ObjectMapper; import org.apache.commons.csv.CSVFormat; import org.apache.commons.csv.CSVPrinter; import org.apache.commons.lang3.mutable.MutableLong; import org.slf4j.Logger; import org.slf4j.LoggerFactory; import org.tallison.cc.index.AbstractRecordProcessor; import org.tallison.cc.index.CCIndexReaderCounter; import org.tallison.cc.index.CCIndexRecord; import org.tallison.cc.index.IndexIterator; import org.apache.tika.exception.TikaConfigException; import org.apache.tika.exception.TikaException; import org.apache.tika.io.TikaInputStream; import org.apache.tika.metadata.Metadata; import org.apache.tika.pipes.FetchEmitTuple; import org.apache.tika.pipes.fetcher.Fetcher; import org.apache.tika.pipes.pipesiterator.CallablePipesIterator; import org.apache.tika.pipes.pipesiterator.PipesIterator; import org.apache.tika.utils.StringUtils; /** * This counts mime_detected. Use a regular file selector to include * only urls that had a 200, e.g. */ public class CCMimeCounter { private static final Long INDEX_WORKER_ID = 1l; private static final Long INDEX_READER_ID = 2l; private static final Logger LOGGER = LoggerFactory.getLogger(CCMimeCounter.class); public static void main(String[] args) throws Exception { ExtractorConfig fetcherConfig = new ObjectMapper().readValue(new File(args[0]), ExtractorConfig.class); execute(fetcherConfig); } private static void execute(ExtractorConfig fetcherConfig) throws IOException, TikaException { ArrayBlockingQueue<FetchEmitTuple> indexPathsList = new ArrayBlockingQueue<>(1000); //IndexPathsReader reads a file containing a list of cc-index.paths files //and writes the literal gz files (cc-index/collections/CC-MAIN-2023-06/indexes/cdx-00000.gz) //to indexPathsList //IndexWorker reads a single index.gz file at a time and processes each record //It fetches non truncated files and logs truncated files int totalThreads = fetcherConfig.getNumThreads() + 1; ExecutorService executorService = Executors.newFixedThreadPool(totalThreads); ExecutorCompletionService<Long> executorCompletionService = new ExecutorCompletionService<>(executorService); IndexIterator indexIterator = fetcherConfig.getIndexIterator(); indexIterator.initialize(Collections.EMPTY_MAP); executorCompletionService.submit(new CallablePipesIterator(indexIterator, indexPathsList)); CCIndexReaderCounter counter = new CCIndexReaderCounter(); int finishedWorkers = 0; List<DetectedMimeCounter> detectedMimeCounters = new ArrayList<>(); try { for (int i = 0; i < fetcherConfig.getNumThreads(); i++) { DetectedMimeCounter processor = new DetectedMimeCounter(fetcherConfig, counter); detectedMimeCounters.add(processor); executorCompletionService.submit( new IndexWorker(fetcherConfig, indexPathsList, processor)); } while (finishedWorkers < totalThreads) { //blocking Future<Long> future = executorCompletionService.take(); if (future != null) { Long f = future.get(); LOGGER.debug("completed {}", f); if (f.equals(INDEX_WORKER_ID)) { finishedWorkers++; } else if (f.equals(INDEX_READER_ID)) { LOGGER.info("Index paths reader successfully completed"); } } } } catch (TikaConfigException e) { LOGGER.error("main loop exception", e); throw new RuntimeException(e); } catch (ExecutionException e) { LOGGER.error("main loop exception", e); throw new RuntimeException(e); } catch (InterruptedException e) { LOGGER.warn("main loop interrupted exception", e); throw new RuntimeException(e); } finally { executorService.shutdown(); executorService.shutdownNow(); } summarize(detectedMimeCounters); } private static void summarize(List<DetectedMimeCounter> detectedMimeCounters) throws IOException { Map<String, Long> total = new HashMap<>(); Map<String, Long> truncated = new HashMap<>(); Map<String, Long> nonTruncated = new HashMap<>(); for (DetectedMimeCounter c : detectedMimeCounters) { update(c.totalCounts, total); update(c.truncatedCounts, truncated); } calcNonTruncated(truncated, total, nonTruncated); report("total", total); report("truncated", truncated); report("non-truncated", nonTruncated); } private static void calcNonTruncated(Map<String, Long> truncated, Map<String, Long> total, Map<String, Long> nonTruncated) { for (Map.Entry<String, Long> e : total.entrySet()) { Long val = e.getValue(); Long t = truncated.getOrDefault(e.getKey(), 0l); val -= t; nonTruncated.put(e.getKey(), val); } } private static void report(String name, Map<String, Long> m) throws IOException { try (BufferedWriter writer = Files.newBufferedWriter(Paths.get(name + ".csv"), StandardCharsets.UTF_8)) { try (CSVPrinter printer = new CSVPrinter(writer, CSVFormat.EXCEL)) { printer.printRecord("mime", "count"); m.entrySet().stream().sorted(Collections.reverseOrder(Map.Entry.comparingByValue())) .forEach(e -> { try { printer.printRecord(e.getKey(), e.getValue()); } catch (IOException ex) { throw new RuntimeException(ex); } }); } } } private static void update(Map<String, MutableLong> from, Map<String, Long> to) { for (Map.Entry<String, MutableLong> e : from.entrySet()) { Long cnt = to.get(e.getKey()); if (cnt == null) { cnt = 0l; } cnt += e.getValue().getValue(); to.put(e.getKey(), cnt); } } private static class IndexWorker implements Callable<Long> { private final ArrayBlockingQueue<FetchEmitTuple> indexUrls; private final AbstractRecordProcessor recordProcessor; private final Fetcher fetcher; IndexWorker(ExtractorConfig fetcherConfig, ArrayBlockingQueue<FetchEmitTuple> indexUrls, AbstractRecordProcessor recordProcessor) throws TikaException { this.indexUrls = indexUrls; this.recordProcessor = recordProcessor; this.fetcher = fetcherConfig.newFetcher(); } @Override public Long call() throws Exception { boolean shouldContinue = true; while (shouldContinue) { FetchEmitTuple indexUrl = indexUrls.poll(120, TimeUnit.MINUTES); if (indexUrl == null) { throw new TimeoutException("waited 120 minutes for a new record"); } if (indexUrl == PipesIterator.COMPLETED_SEMAPHORE) { recordProcessor.close(); //can hang forever indexUrls.put(PipesIterator.COMPLETED_SEMAPHORE); return INDEX_WORKER_ID; } shouldContinue = processFile(indexUrl, recordProcessor); } return INDEX_WORKER_ID; } private boolean processFile(FetchEmitTuple fetchEmitTuple, AbstractRecordProcessor recordProcessor) throws InterruptedException { long start = System.currentTimeMillis(); LOGGER.info("starting to fetch index gz: {}", fetchEmitTuple.getFetchKey().getFetchKey()); try (TikaInputStream tis = (TikaInputStream) fetcher.fetch( fetchEmitTuple.getFetchKey().getFetchKey(), new Metadata())) { try (InputStream is = new BufferedInputStream(new GZIPInputStream(tis))) { try (BufferedReader reader = new BufferedReader( new InputStreamReader(is, StandardCharsets.UTF_8))) { String line = reader.readLine(); int lines = 0; long elapsed = System.currentTimeMillis() - start; LOGGER.info("Finished fetching {} bytes in {} ms for index gz: {}", String.format(Locale.US, "%,d", tis.getLength()), String.format(Locale.US, "%,d", elapsed), fetchEmitTuple.getFetchKey().getFetchKey()); while (line != null) { LOGGER.trace("about to add a line"); if (StringUtils.isBlank(line)) { line = reader.readLine(); continue; } try { boolean shouldContinue = recordProcessor.process(line); if (!shouldContinue) { return shouldContinue; } } catch (IOException e) { LOGGER.warn("bad json: " + line); } lines++; line = reader.readLine(); } } } } catch (TikaException | IOException e) { LOGGER.error( "failed while processing " + fetchEmitTuple.getFetchKey().getFetchKey(), e); } long elapsed = System.currentTimeMillis() - start; LOGGER.info("finished processing index gz in ({}) ms: {}", String.format(Locale.US, "%,d", elapsed), fetchEmitTuple.getFetchKey().getFetchKey()); return true; } } private static class DetectedMimeCounter extends AbstractRecordProcessor { private final ExtractorConfig fetcherConfig; private final CCIndexReaderCounter counter; private final Map<String, MutableLong> totalCounts = new HashMap<>(); private final Map<String, MutableLong> truncatedCounts = new HashMap<>(); public DetectedMimeCounter(ExtractorConfig fetcherConfig, CCIndexReaderCounter counter) { this.fetcherConfig = fetcherConfig; this.counter = counter; } @Override public boolean process(String json) throws IOException, InterruptedException { long totalRead = counter.getRecordsRead().incrementAndGet(); if (totalRead % 1000000 == 0) { LOGGER.info("processed: {}", counter); } if (fetcherConfig.getMaxRecords() > -1 && totalRead >= fetcherConfig.getMaxRecords()) { LOGGER.info("hit max read"); return false; } //check for hit max //return false; Optional<CCIndexRecord> record = CCIndexRecord.parseRecord(json); if (record.isEmpty()) { //problem already logged return true; } CCIndexRecord r = record.get(); if (!fetcherConfig.getRecordSelector().select(r)) { return true; } increment(totalCounts, r.getNormalizedMimeDetected()); if (!StringUtils.isBlank(r.getTruncated())) { long truncated = counter.getTruncated().incrementAndGet(); if (fetcherConfig.getMaxFilesTruncated() > -1 && truncated >= fetcherConfig.getMaxFilesTruncated()) { LOGGER.info("hit max truncated files"); return false; } increment(truncatedCounts,
r.getNormalizedMimeDetected());
return true; } return true; } private void increment(Map<String, MutableLong> m, String k) { MutableLong cnt = m.get(k); if (cnt == null) { cnt = new MutableLong(1); m.put(k, cnt); return; } else { cnt.increment(); } } @Override public void close() throws IOException { } } }
src/main/java/org/tallison/cc/index/extractor/CCMimeCounter.java
tballison-commoncrawl-fetcher-lite-2c106d1
[ { "filename": "src/main/java/org/tallison/cc/index/extractor/CCFileExtractorRecordProcessor.java", "retrieved_chunk": " if (fetcherConfig.getMaxFilesTruncated() > -1 &&\n truncated >= fetcherConfig.getMaxFilesTruncated()) {\n LOGGER.info(\"hit max truncated files\");\n return false;\n }\n }\n if (fetcherConfig.isExtractTruncated() || StringUtils.isBlank(r.getTruncated())) {\n long extracted = counter.getFilesExtracted().incrementAndGet();\n if (fetcherConfig.getMaxFilesExtracted() > -1 &&\n extracted >= fetcherConfig.getMaxFilesExtracted()) {", "score": 0.9271911382675171 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/CCFileExtractorRecordProcessor.java", "retrieved_chunk": " LOGGER.info(\"hit max extracted files\");\n return false;\n }\n if (fetcherConfig.isDryRun()) {\n LOGGER.info(\"dry run, but would have extracted {}\", r);\n return true;\n }\n fetchBytes(r);\n return true;\n } else {", "score": 0.8687166571617126 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/CCFileExtractorRecordProcessor.java", "retrieved_chunk": " //problem already logged\n return true;\n }\n CCIndexRecord r = record.get();\n if (!fetcherConfig.getRecordSelector().select(r)) {\n return true;\n }\n //if truncated, count appropriately and test for limits\n if (!StringUtils.isBlank(r.getTruncated())) {\n long truncated = counter.getTruncated().incrementAndGet();", "score": 0.8622099757194519 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/CCFileExtractorRecordProcessor.java", "retrieved_chunk": " String url = r.getUrl();\n TRUNCATED_URLS_LOGGER.info(\"\", url);\n //url,mime,mime_detected,warc_file,warc_offset,warc_length,truncated\n TRUNCATED_URLS_FULL_LOGGER.info(\"\", url,\n r.getNormalizedMime(), r.getNormalizedMimeDetected(), r.getFilename(),\n r.getOffset(), r.getLength(), r.getTruncated());\n return true;\n }\n }\n private void fetchBytes(CCIndexRecord r) throws InterruptedException {", "score": 0.8427147269248962 }, { "filename": "src/main/java/org/tallison/cc/index/extractor/CCFileExtractorRecordProcessor.java", "retrieved_chunk": " LOGGER.info(\"processed: {}\", counter);\n }\n if (fetcherConfig.getMaxRecords() > -1 && totalRead >= fetcherConfig.getMaxRecords()) {\n LOGGER.info(\"hit max read\");\n return false;\n }\n //check for hit max\n //return false;\n Optional<CCIndexRecord> record = CCIndexRecord.parseRecord(json);\n if (record.isEmpty()) {", "score": 0.8286970853805542 } ]
java
r.getNormalizedMimeDetected());
/* * Copyright © 2023 Deutsche Telekom AG (opensource@telekom.de) * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ package de.telekom.phonenumbernormalizer.numberplans; import org.slf4j.Logger; import org.slf4j.LoggerFactory; import de.telekom.phonenumbernormalizer.dto.DeviceContextLineType; import de.telekom.phonenumbernormalizer.numberplans.constants.DeFixedLineNumberPlan; import java.util.EnumMap; import java.util.HashMap; import java.util.Map; /** * A factory to retrieve a NumberPlan for a given line-type and country calling code. Currently supporting: * <ul> * <li>German Fixed-Line</li> * </ul> * * @see NumberPlanFactory#getNumberPlan(DeviceContextLineType, String) */ public class NumberPlanFactory { private static final Logger LOGGER = LoggerFactory.getLogger(NumberPlanFactory.class); public static final NumberPlanFactory INSTANCE = new NumberPlanFactory(); /** * Two-dimensional map - The first key is DeviceContextLineType and second key is the Country Calling Code while the value is a NumberPlan object. * * @see NumberPlan * @see DeviceContextLineType */ private final Map<DeviceContextLineType, Map<String, NumberPlan>> numberPlans = new EnumMap<>(DeviceContextLineType.class); /** * Adding all coded NumberPlans to the factory */ private NumberPlanFactory() { this.initFixedLineNumberPlans(); this.initMobileNumberPlans(); this.initFallBackNumberPlans(); } /** * Adds coded NumberPlans for fixed-line context to the factory. * * @see NumberPlan * @see DeviceContextLineType#FIXEDLINE */ private void initFixedLineNumberPlans() { Map<String, NumberPlan> fixedLineNumberPlans = new HashMap<>(); fixedLineNumberPlans
.put(DeFixedLineNumberPlan.getCountryCode(), new DeFixedLineNumberPlan());
numberPlans.put(DeviceContextLineType.FIXEDLINE, fixedLineNumberPlans); } /** * Adds coded NumberPlans for mobile context to the factory. * * @see NumberPlan * @see DeviceContextLineType#MOBILE */ private void initMobileNumberPlans() { // TODO: Mobile Number Plan } /** * Adds coded NumberPlans for unknown context to the factory. These are just the common rules for mobile and fixed-line, so that they could be applied regardless of which actual line context is used. * * @see NumberPlan * @see DeviceContextLineType#UNKNOWN */ private void initFallBackNumberPlans() { Map<String, NumberPlan> fixedLineNumberPlans = new HashMap<>(); // For Germany all short numbers of the fixed-line are also valid in mobile, so we can reuse it, if unknown. fixedLineNumberPlans.put(DeFixedLineNumberPlan.getCountryCode(), new DeFixedLineNumberPlan()); numberPlans.put(DeviceContextLineType.UNKNOWN, fixedLineNumberPlans); } /** * Gets a NumberPlan for a line-type of a specific country. * * @param numberPlanType line-type where the NumberPlan is valid * @param countryCode country calling code for which the NumberPlan * @return {@link NumberPlan} for further checks * * @see DeviceContextLineType * @see NumberPlan#isMatchingShortNumber(String) */ public NumberPlan getNumberPlan(DeviceContextLineType numberPlanType, String countryCode) { if (numberPlans.containsKey(numberPlanType)) { LOGGER.debug("use number plan for type: {}", numberPlanType); Map<String, NumberPlan> numberPlan = numberPlans.get(numberPlanType); if (numberPlan.containsKey(countryCode)) { LOGGER.debug("use number plan for country code: {}", countryCode); return numberPlan.get(countryCode); } } LOGGER.debug("no number plan for country available"); return null; } }
src/main/java/de/telekom/phonenumbernormalizer/numberplans/NumberPlanFactory.java
telekom-phonenumber-normalizer-9a1b7ad
[ { "filename": "src/main/java/de/telekom/phonenumbernormalizer/numberplans/constants/DeFixedLineNumberPlan.java", "retrieved_chunk": " *\n */\npublic class DeFixedLineNumberPlan extends NumberPlan {\n /**\n * Constant for German Country Calling Code\n */\n private static final String COUNTRY_CODE = \"49\";\n /**\n * Constant for German short numbers in fixed-line\n */", "score": 0.864551842212677 }, { "filename": "src/main/java/de/telekom/phonenumbernormalizer/numberplans/NumberPlan.java", "retrieved_chunk": " * Additionally, we designed the NumberPlanFactory to have a NumberPlan class for each DeviceContextLineType, so we can support short numbers which are valid only in fixed-line or mobile context.\n * </p>\n * @see NumberPlanFactory\n * @see de.telekom.phonenumbernormalizer.dto.DeviceContextLineType\n */\npublic abstract class NumberPlan {\n private static final Logger LOGGER = LoggerFactory.getLogger(NumberPlan.class);\n /**\n * A subclass needs to provide a Map&lt;String, Integer&gt; as tbe rules to identify short numbers\n * The key (String) is representing a prefix for the number and the value (Integer) is the total length of the short code (including the prefix)", "score": 0.8387374877929688 }, { "filename": "src/main/java/de/telekom/phonenumbernormalizer/PhoneNumberNormalizerImpl.java", "retrieved_chunk": " NumberPlan numberplan = null;\n if (deviceContext != null) {\n numberplan = NumberPlanFactory.INSTANCE.getNumberPlan(deviceContext.getLineType(), deviceContext.getCountryCode());\n }\n if (wrapper.isShortNumber(numberplan)) {\n //if it is a short number, we can't add area code nor country code, so returning the dialable.\n return wrapper.getDialableNumber();\n }\n if (wrapper.hasRegionNationalAccessCode() && deviceContext != null) {\n //Number plan is using a NationalPrefix aka Trunc Code ... so we could add Area Code if not included in the number.", "score": 0.8222073316574097 }, { "filename": "src/main/java/de/telekom/phonenumbernormalizer/numberplans/constants/DeFixedLineNumberPlan.java", "retrieved_chunk": " * distributed under the License is distributed on an \"AS IS\" BASIS,\n * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n * See the License for the specific language governing permissions and\n * limitations under the License.\n */\npackage de.telekom.phonenumbernormalizer.numberplans.constants;\nimport java.util.Map;\nimport de.telekom.phonenumbernormalizer.numberplans.NumberPlan;\n/**\n * Definition see Chapter 8.1 in <a href=\"https://www.bundesnetzagentur.de/SharedDocs/Downloads/DE/Sachgebiete/Telekommunikation/Unternehmen_Institutionen/Nummerierung/Nummerierungskonzept/Nummerierungskonzept2011pdf.pdf?__blob=publicationFile\">BNetzA German Number Plan</a>", "score": 0.8112590312957764 }, { "filename": "src/main/java/de/telekom/phonenumbernormalizer/PhoneNumberNormalizerImpl.java", "retrieved_chunk": " * distributed under the License is distributed on an \"AS IS\" BASIS,\n * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n * See the License for the specific language governing permissions and\n * limitations under the License.\n */\npackage de.telekom.phonenumbernormalizer;\nimport de.telekom.phonenumbernormalizer.dto.DeviceContext;\nimport de.telekom.phonenumbernormalizer.dto.DeviceContextMapper;\nimport de.telekom.phonenumbernormalizer.numberplans.PhoneLibWrapper;\nimport de.telekom.phonenumbernormalizer.numberplans.NumberPlan;", "score": 0.8104901313781738 } ]
java
.put(DeFixedLineNumberPlan.getCountryCode(), new DeFixedLineNumberPlan());
/* * Copyright © 2023 Deutsche Telekom AG (opensource@telekom.de) * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ package de.telekom.phonenumbernormalizer; import de.telekom.phonenumbernormalizer.dto.DeviceContext; import de.telekom.phonenumbernormalizer.dto.DeviceContextMapper; import de.telekom.phonenumbernormalizer.numberplans.PhoneLibWrapper; import de.telekom.phonenumbernormalizer.numberplans.NumberPlan; import de.telekom.phonenumbernormalizer.numberplans.NumberPlanFactory; import lombok.RequiredArgsConstructor; import org.slf4j.Logger; import org.slf4j.LoggerFactory; import org.springframework.stereotype.Component; /** * Concrete implementation of {@link PhoneNumberNormalizer} using {@link PhoneLibWrapper} to normalize a number by mitigating some inaccuracies when it comes to number plans of optional NDC and NAC as zero. * <p> * Also supports {@link DeviceContext} to enrich a phone number during normalization if the optional NDC is missing. * </p> */ @RequiredArgsConstructor @Component public class PhoneNumberNormalizerImpl implements PhoneNumberNormalizer { private static final Logger LOGGER = LoggerFactory.getLogger(PhoneNumberNormalizerImpl.class); /** * Storage for {@link PhoneNumberNormalizer#setFallbackRegionCode(String)} */ private String fallbackRegionCode = null; @Override public void setFallbackRegionCode(String fallBackRegionCode) { if (PhoneLibWrapper.getCountryCodeForRegion(fallBackRegionCode) > 0) { this.fallbackRegionCode = fallBackRegionCode; } else { this.fallbackRegionCode = null; //invalid region code! } } /** * Fallback normalization within the number-plan of the fallback region. * @param number the original number to be normalized * @param dialableNumber the original number reduced to dialable digits * @return E164 formatted phone number or at least a dialable version of the number * * @see PhoneNumberNormalizer#setFallbackRegionCode(String) * @see PhoneNumberNormalizer#normalizePhoneNumber(String, String) */ private String fallbackNormalizationFromDeviceContextToDefaultRegionCode(String number, String dialableNumber) { if (this.fallbackRegionCode == null) { LOGGER.debug("Fallback Region was set!"); return dialableNumber; } else { return this.normalizePhoneNumber(number, this.fallbackRegionCode); } } /** * Uses wrapper of PhoneLib to identify if special rules apply for normalization.<br/> * Using device context for enriching the number make it normalizable to E164 format if NDC is optional in the used number plan, but not used in the phone number to be normalized. * @param wrapper instanced wrapper of PhoneLib * @param deviceContext information like CC, NDC and {@link de.telekom.phonenumbernormalizer.dto.DeviceContextLineType} from which the number is dialled * @return E164 formatted phone number or dialable version of it or null */ private String normalize(PhoneLibWrapper wrapper, DeviceContext deviceContext) { // international prefix has been added by PhoneLib even if it's not valid in the number plan. if (wrapper == null) { LOGGER.debug("PhoneLipWrapper was not initialized"); return null; } if (wrapper.getSemiNormalizedNumber() == null) { return wrapper.getDialableNumber(); } NumberPlan numberplan = null; if (deviceContext != null) { numberplan = NumberPlanFactory.INSTANCE.getNumberPlan(deviceContext.getLineType
(), deviceContext.getCountryCode());
} if (wrapper.isShortNumber(numberplan)) { //if it is a short number, we can't add area code nor country code, so returning the dialable. return wrapper.getDialableNumber(); } if (wrapper.hasRegionNationalAccessCode() && deviceContext != null) { //Number plan is using a NationalPrefix aka Trunc Code ... so we could add Area Code if not included in the number. return wrapper.extendNumberByDefaultAreaCodeAndCountryCode(wrapper.getNationalAccessCode(), deviceContext.getNationalDestinationCode()); } // Number plan is not using NationalPrefix aka Trunc Code ... its also not a short number, so country code can be added: return wrapper.getE164Formatted(); } @Override public String normalizePhoneNumber(String number, String regionCode) { PhoneLibWrapper wrapper = new PhoneLibWrapper(number, regionCode); if (wrapper.getSemiNormalizedNumber() == null) { return wrapper.getDialableNumber(); } if (wrapper.isShortNumber()) { //if it is a short number, we can't add area code nor country code, so returning the dialable. return wrapper.getDialableNumber(); } // international prefix is added by the lib even if it's not valid in the number plan. //checking if the input number is equal to the nationalNumber based on number plan and trunc code logic. boolean hasNoCCAndNoNAC = wrapper.hasNoCountryCodeNorNationalAccessCode(); LOGGER.debug("Number has no CC and no NAC: {}.", hasNoCCAndNoNAC); //if the number is definitely a short number or needs an area code but does not have it, we do not add the country code. return (hasNoCCAndNoNAC) ? wrapper.getDialableNumber() : wrapper.getE164Formatted(); } @Override public String normalizePhoneNumber(String number, DeviceContext deviceContext) { // checking if the number has a special format or is not valid at all. PhoneLibWrapper normalizerPhoneNumber = new PhoneLibWrapper(number, null); if (! normalizerPhoneNumber.isNormalizingTried()) { return normalizerPhoneNumber.getDialableNumber(); } DeviceContext normalizedDeviceContext = DeviceContextMapper.normalized(deviceContext); if (!normalizedDeviceContext.getCountryCode().equals(DeviceContext.UNKNOWN_VALUE)) { String regionCode = PhoneLibWrapper.getRegionCodeForCountryCode(normalizedDeviceContext.getCountryCode()); // now working again with the region code normalizerPhoneNumber = new PhoneLibWrapper(number, regionCode); if (!normalizedDeviceContext.getNationalDestinationCode().equals(DeviceContext.UNKNOWN_VALUE)) { // Number needs normalization: return normalize(normalizerPhoneNumber, deviceContext); } // Device Context with CountryCode but without AreaCode ... if (!(PhoneLibWrapper.UNKNOWN_REGIONCODE.equals(regionCode))) { return this.normalizePhoneNumber(number, regionCode); } } LOGGER.debug("Normalization based on DeviceContext did not work - falling back to normalization with fallback region."); return this.fallbackNormalizationFromDeviceContextToDefaultRegionCode(number, normalizerPhoneNumber.getDialableNumber()); } }
src/main/java/de/telekom/phonenumbernormalizer/PhoneNumberNormalizerImpl.java
telekom-phonenumber-normalizer-9a1b7ad
[ { "filename": "src/main/java/de/telekom/phonenumbernormalizer/numberplans/PhoneLibWrapper.java", "retrieved_chunk": " } else {\n if (!isSpecialFormat(dialableNumber)) {\n // Number needs normalization:\n // international prefix is added by the lib even if it's not valid in the number plan.\n this.isNormalizingTried = true;\n this.semiNormalizedNumber = PhoneLibWrapper.parseNumber(dialableNumber, regionCode);\n }\n }\n }\n }", "score": 0.8675501346588135 }, { "filename": "src/main/java/de/telekom/phonenumbernormalizer/numberplans/PhoneLibWrapper.java", "retrieved_chunk": " String dialableNumber;\n /**\n * The given number normalized with PhoneLib, risking we get a incorrect normalization\n *\n * @see PhoneLibWrapper#PhoneLibWrapper(String, String)\n * @see PhoneLibWrapper#isNormalizingTried()\n * @see PhoneLibWrapper#getSemiNormalizedNumber()\n */\n Phonenumber.PhoneNumber semiNormalizedNumber;\n /**", "score": 0.8560713529586792 }, { "filename": "src/main/java/de/telekom/phonenumbernormalizer/numberplans/PhoneLibWrapper.java", "retrieved_chunk": " try {\n this.semiNormalizedNumber = phoneUtil.parse(extendedNumber, regionCode);\n // after area code has been added, we can add the country code by the lib:\n return getE164Formatted();\n } catch (NumberParseException e) {\n LOGGER.warn(\"could not parse extended number: {}\", extendedNumber);\n LOGGER.debug(\"{}\", e.getMessage());\n return dialableNumber;\n }\n } else {", "score": 0.8556559681892395 }, { "filename": "src/main/java/de/telekom/phonenumbernormalizer/numberplans/NumberPlanFactory.java", "retrieved_chunk": " LOGGER.debug(\"no number plan for country available\");\n return null;\n }\n}", "score": 0.8546358346939087 }, { "filename": "src/main/java/de/telekom/phonenumbernormalizer/numberplans/NumberPlanFactory.java", "retrieved_chunk": " */\n public NumberPlan getNumberPlan(DeviceContextLineType numberPlanType, String countryCode) {\n if (numberPlans.containsKey(numberPlanType)) {\n LOGGER.debug(\"use number plan for type: {}\", numberPlanType);\n Map<String, NumberPlan> numberPlan = numberPlans.get(numberPlanType);\n if (numberPlan.containsKey(countryCode)) {\n LOGGER.debug(\"use number plan for country code: {}\", countryCode);\n return numberPlan.get(countryCode);\n }\n }", "score": 0.8497321605682373 } ]
java
(), deviceContext.getCountryCode());
/* * Copyright © 2023 Deutsche Telekom AG (opensource@telekom.de) * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ package de.telekom.phonenumbernormalizer; import de.telekom.phonenumbernormalizer.dto.DeviceContext; import de.telekom.phonenumbernormalizer.dto.DeviceContextMapper; import de.telekom.phonenumbernormalizer.numberplans.PhoneLibWrapper; import de.telekom.phonenumbernormalizer.numberplans.NumberPlan; import de.telekom.phonenumbernormalizer.numberplans.NumberPlanFactory; import lombok.RequiredArgsConstructor; import org.slf4j.Logger; import org.slf4j.LoggerFactory; import org.springframework.stereotype.Component; /** * Concrete implementation of {@link PhoneNumberNormalizer} using {@link PhoneLibWrapper} to normalize a number by mitigating some inaccuracies when it comes to number plans of optional NDC and NAC as zero. * <p> * Also supports {@link DeviceContext} to enrich a phone number during normalization if the optional NDC is missing. * </p> */ @RequiredArgsConstructor @Component public class PhoneNumberNormalizerImpl implements PhoneNumberNormalizer { private static final Logger LOGGER = LoggerFactory.getLogger(PhoneNumberNormalizerImpl.class); /** * Storage for {@link PhoneNumberNormalizer#setFallbackRegionCode(String)} */ private String fallbackRegionCode = null; @Override public void setFallbackRegionCode(String fallBackRegionCode) { if (PhoneLibWrapper.getCountryCodeForRegion(fallBackRegionCode) > 0) { this.fallbackRegionCode = fallBackRegionCode; } else { this.fallbackRegionCode = null; //invalid region code! } } /** * Fallback normalization within the number-plan of the fallback region. * @param number the original number to be normalized * @param dialableNumber the original number reduced to dialable digits * @return E164 formatted phone number or at least a dialable version of the number * * @see PhoneNumberNormalizer#setFallbackRegionCode(String) * @see PhoneNumberNormalizer#normalizePhoneNumber(String, String) */ private String fallbackNormalizationFromDeviceContextToDefaultRegionCode(String number, String dialableNumber) { if (this.fallbackRegionCode == null) { LOGGER.debug("Fallback Region was set!"); return dialableNumber; } else { return this.normalizePhoneNumber(number, this.fallbackRegionCode); } } /** * Uses wrapper of PhoneLib to identify if special rules apply for normalization.<br/> * Using device context for enriching the number make it normalizable to E164 format if NDC is optional in the used number plan, but not used in the phone number to be normalized. * @param wrapper instanced wrapper of PhoneLib * @param deviceContext information like CC, NDC and {@link de.telekom.phonenumbernormalizer.dto.DeviceContextLineType} from which the number is dialled * @return E164 formatted phone number or dialable version of it or null */ private String normalize(PhoneLibWrapper wrapper, DeviceContext deviceContext) { // international prefix has been added by PhoneLib even if it's not valid in the number plan. if (wrapper == null) { LOGGER.debug("PhoneLipWrapper was not initialized"); return null; } if (wrapper.getSemiNormalizedNumber() == null) { return wrapper.getDialableNumber(); } NumberPlan numberplan = null; if (deviceContext != null) { numberplan = NumberPlanFactory.INSTANCE.getNumberPlan(deviceContext.getLineType(), deviceContext.getCountryCode()); } if (wrapper.isShortNumber(numberplan)) { //if it is a short number, we can't add area code nor country code, so returning the dialable. return wrapper.getDialableNumber(); } if (wrapper.hasRegionNationalAccessCode() && deviceContext != null) { //Number plan is using a NationalPrefix aka Trunc Code ... so we could add Area Code if not included in the number. return wrapper.extendNumberByDefaultAreaCodeAndCountryCode(wrapper.getNationalAccessCode(),
deviceContext.getNationalDestinationCode());
} // Number plan is not using NationalPrefix aka Trunc Code ... its also not a short number, so country code can be added: return wrapper.getE164Formatted(); } @Override public String normalizePhoneNumber(String number, String regionCode) { PhoneLibWrapper wrapper = new PhoneLibWrapper(number, regionCode); if (wrapper.getSemiNormalizedNumber() == null) { return wrapper.getDialableNumber(); } if (wrapper.isShortNumber()) { //if it is a short number, we can't add area code nor country code, so returning the dialable. return wrapper.getDialableNumber(); } // international prefix is added by the lib even if it's not valid in the number plan. //checking if the input number is equal to the nationalNumber based on number plan and trunc code logic. boolean hasNoCCAndNoNAC = wrapper.hasNoCountryCodeNorNationalAccessCode(); LOGGER.debug("Number has no CC and no NAC: {}.", hasNoCCAndNoNAC); //if the number is definitely a short number or needs an area code but does not have it, we do not add the country code. return (hasNoCCAndNoNAC) ? wrapper.getDialableNumber() : wrapper.getE164Formatted(); } @Override public String normalizePhoneNumber(String number, DeviceContext deviceContext) { // checking if the number has a special format or is not valid at all. PhoneLibWrapper normalizerPhoneNumber = new PhoneLibWrapper(number, null); if (! normalizerPhoneNumber.isNormalizingTried()) { return normalizerPhoneNumber.getDialableNumber(); } DeviceContext normalizedDeviceContext = DeviceContextMapper.normalized(deviceContext); if (!normalizedDeviceContext.getCountryCode().equals(DeviceContext.UNKNOWN_VALUE)) { String regionCode = PhoneLibWrapper.getRegionCodeForCountryCode(normalizedDeviceContext.getCountryCode()); // now working again with the region code normalizerPhoneNumber = new PhoneLibWrapper(number, regionCode); if (!normalizedDeviceContext.getNationalDestinationCode().equals(DeviceContext.UNKNOWN_VALUE)) { // Number needs normalization: return normalize(normalizerPhoneNumber, deviceContext); } // Device Context with CountryCode but without AreaCode ... if (!(PhoneLibWrapper.UNKNOWN_REGIONCODE.equals(regionCode))) { return this.normalizePhoneNumber(number, regionCode); } } LOGGER.debug("Normalization based on DeviceContext did not work - falling back to normalization with fallback region."); return this.fallbackNormalizationFromDeviceContextToDefaultRegionCode(number, normalizerPhoneNumber.getDialableNumber()); } }
src/main/java/de/telekom/phonenumbernormalizer/PhoneNumberNormalizerImpl.java
telekom-phonenumber-normalizer-9a1b7ad
[ { "filename": "src/main/java/de/telekom/phonenumbernormalizer/numberplans/PhoneLibWrapper.java", "retrieved_chunk": " } else {\n if (!isSpecialFormat(dialableNumber)) {\n // Number needs normalization:\n // international prefix is added by the lib even if it's not valid in the number plan.\n this.isNormalizingTried = true;\n this.semiNormalizedNumber = PhoneLibWrapper.parseNumber(dialableNumber, regionCode);\n }\n }\n }\n }", "score": 0.8710631132125854 }, { "filename": "src/main/java/de/telekom/phonenumbernormalizer/numberplans/PhoneLibWrapper.java", "retrieved_chunk": " try {\n this.semiNormalizedNumber = phoneUtil.parse(extendedNumber, regionCode);\n // after area code has been added, we can add the country code by the lib:\n return getE164Formatted();\n } catch (NumberParseException e) {\n LOGGER.warn(\"could not parse extended number: {}\", extendedNumber);\n LOGGER.debug(\"{}\", e.getMessage());\n return dialableNumber;\n }\n } else {", "score": 0.8672100901603699 }, { "filename": "src/main/java/de/telekom/phonenumbernormalizer/numberplans/NumberPlanFactory.java", "retrieved_chunk": " */\n public NumberPlan getNumberPlan(DeviceContextLineType numberPlanType, String countryCode) {\n if (numberPlans.containsKey(numberPlanType)) {\n LOGGER.debug(\"use number plan for type: {}\", numberPlanType);\n Map<String, NumberPlan> numberPlan = numberPlans.get(numberPlanType);\n if (numberPlan.containsKey(countryCode)) {\n LOGGER.debug(\"use number plan for country code: {}\", countryCode);\n return numberPlan.get(countryCode);\n }\n }", "score": 0.8588719367980957 }, { "filename": "src/main/java/de/telekom/phonenumbernormalizer/numberplans/PhoneLibWrapper.java", "retrieved_chunk": " * @param defaultNationalDestinationCode the NDC to be added depending on the use telephone line origination.\n * @return if possible a E164 formatted representation or just the diallable representation of the given number.\n *\n * @see PhoneLibWrapper#PhoneLibWrapper(String, String)\n */\n public String extendNumberByDefaultAreaCodeAndCountryCode(String nationalAccessCode, String defaultNationalDestinationCode) {\n String nationalPhoneNumberWithoutNationalAccessCode = this.getNationalPhoneNumberWithoutNationalAccessCode();\n //if the dialableNumber is same as the national Number, Without NationalPrefix, then there is no NDC, so it needs to be added.\n if (Objects.equals(dialableNumber, nationalPhoneNumberWithoutNationalAccessCode)) {\n String extendedNumber = nationalAccessCode + defaultNationalDestinationCode + nationalPhoneNumberWithoutNationalAccessCode;", "score": 0.8576503992080688 }, { "filename": "src/main/java/de/telekom/phonenumbernormalizer/numberplans/PhoneLibWrapper.java", "retrieved_chunk": " * If we have a plain national number based on regions number plan and potential NAC logic.\n * <p>\n * For a number plan without NAC logic, it will always return false!\n * </p>\n * @return if given number could have CC and NAC, but does not have any of them.\n */\n public boolean hasNoCountryCodeNorNationalAccessCode() {\n // if given number has no NAC and no CC, it equals national phone number (without NAC).\n if (! Objects.equals(dialableNumber, this.getNationalPhoneNumberWithoutNationalAccessCode())) {\n return false;", "score": 0.8510130047798157 } ]
java
deviceContext.getNationalDestinationCode());
/* * Copyright © 2023 Deutsche Telekom AG (opensource@telekom.de) * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ package de.telekom.phonenumbernormalizer; import de.telekom.phonenumbernormalizer.dto.DeviceContext; import de.telekom.phonenumbernormalizer.dto.DeviceContextMapper; import de.telekom.phonenumbernormalizer.numberplans.PhoneLibWrapper; import de.telekom.phonenumbernormalizer.numberplans.NumberPlan; import de.telekom.phonenumbernormalizer.numberplans.NumberPlanFactory; import lombok.RequiredArgsConstructor; import org.slf4j.Logger; import org.slf4j.LoggerFactory; import org.springframework.stereotype.Component; /** * Concrete implementation of {@link PhoneNumberNormalizer} using {@link PhoneLibWrapper} to normalize a number by mitigating some inaccuracies when it comes to number plans of optional NDC and NAC as zero. * <p> * Also supports {@link DeviceContext} to enrich a phone number during normalization if the optional NDC is missing. * </p> */ @RequiredArgsConstructor @Component public class PhoneNumberNormalizerImpl implements PhoneNumberNormalizer { private static final Logger LOGGER = LoggerFactory.getLogger(PhoneNumberNormalizerImpl.class); /** * Storage for {@link PhoneNumberNormalizer#setFallbackRegionCode(String)} */ private String fallbackRegionCode = null; @Override public void setFallbackRegionCode(String fallBackRegionCode) { if (PhoneLibWrapper.getCountryCodeForRegion(fallBackRegionCode) > 0) { this.fallbackRegionCode = fallBackRegionCode; } else { this.fallbackRegionCode = null; //invalid region code! } } /** * Fallback normalization within the number-plan of the fallback region. * @param number the original number to be normalized * @param dialableNumber the original number reduced to dialable digits * @return E164 formatted phone number or at least a dialable version of the number * * @see PhoneNumberNormalizer#setFallbackRegionCode(String) * @see PhoneNumberNormalizer#normalizePhoneNumber(String, String) */ private String fallbackNormalizationFromDeviceContextToDefaultRegionCode(String number, String dialableNumber) { if (this.fallbackRegionCode == null) { LOGGER.debug("Fallback Region was set!"); return dialableNumber; } else { return this.normalizePhoneNumber(number, this.fallbackRegionCode); } } /** * Uses wrapper of PhoneLib to identify if special rules apply for normalization.<br/> * Using device context for enriching the number make it normalizable to E164 format if NDC is optional in the used number plan, but not used in the phone number to be normalized. * @param wrapper instanced wrapper of PhoneLib * @param deviceContext information like CC, NDC and {@link de.telekom.phonenumbernormalizer.dto.DeviceContextLineType} from which the number is dialled * @return E164 formatted phone number or dialable version of it or null */ private String normalize(PhoneLibWrapper wrapper, DeviceContext deviceContext) { // international prefix has been added by PhoneLib even if it's not valid in the number plan. if (wrapper == null) { LOGGER.debug("PhoneLipWrapper was not initialized"); return null; } if (wrapper.getSemiNormalizedNumber() == null) { return wrapper.getDialableNumber(); } NumberPlan numberplan = null; if (deviceContext != null) { numberplan = NumberPlanFactory.INSTANCE.getNumberPlan(deviceContext.getLineType(), deviceContext.getCountryCode()); } if (wrapper.isShortNumber(numberplan)) { //if it is a short number, we can't add area code nor country code, so returning the dialable. return wrapper.getDialableNumber(); } if (wrapper.hasRegionNationalAccessCode() && deviceContext != null) { //Number plan is using a NationalPrefix aka Trunc Code ... so we could add Area Code if not included in the number. return wrapper.extendNumberByDefaultAreaCodeAndCountryCode(wrapper.getNationalAccessCode(), deviceContext.getNationalDestinationCode()); } // Number plan is not using NationalPrefix aka Trunc Code ... its also not a short number, so country code can be added: return wrapper.getE164Formatted(); } @Override public String normalizePhoneNumber(String number, String regionCode) { PhoneLibWrapper wrapper = new PhoneLibWrapper(number, regionCode); if (wrapper.getSemiNormalizedNumber() == null) { return wrapper.getDialableNumber(); } if (wrapper.isShortNumber()) { //if it is a short number, we can't add area code nor country code, so returning the dialable. return wrapper.getDialableNumber(); } // international prefix is added by the lib even if it's not valid in the number plan. //checking if the input number is equal to the nationalNumber based on number plan and trunc code logic. boolean hasNoCCAndNoNAC = wrapper.hasNoCountryCodeNorNationalAccessCode(); LOGGER.debug("Number has no CC and no NAC: {}.", hasNoCCAndNoNAC); //if the number is definitely a short number or needs an area code but does not have it, we do not add the country code. return (hasNoCCAndNoNAC) ? wrapper.getDialableNumber() : wrapper.getE164Formatted(); } @Override public String normalizePhoneNumber(String number, DeviceContext deviceContext) { // checking if the number has a special format or is not valid at all. PhoneLibWrapper normalizerPhoneNumber = new PhoneLibWrapper(number, null); if (! normalizerPhoneNumber.isNormalizingTried()) { return normalizerPhoneNumber.getDialableNumber(); } DeviceContext normalizedDeviceContext = DeviceContextMapper.normalized(deviceContext); if
(!normalizedDeviceContext.getCountryCode().equals(DeviceContext.UNKNOWN_VALUE)) {
String regionCode = PhoneLibWrapper.getRegionCodeForCountryCode(normalizedDeviceContext.getCountryCode()); // now working again with the region code normalizerPhoneNumber = new PhoneLibWrapper(number, regionCode); if (!normalizedDeviceContext.getNationalDestinationCode().equals(DeviceContext.UNKNOWN_VALUE)) { // Number needs normalization: return normalize(normalizerPhoneNumber, deviceContext); } // Device Context with CountryCode but without AreaCode ... if (!(PhoneLibWrapper.UNKNOWN_REGIONCODE.equals(regionCode))) { return this.normalizePhoneNumber(number, regionCode); } } LOGGER.debug("Normalization based on DeviceContext did not work - falling back to normalization with fallback region."); return this.fallbackNormalizationFromDeviceContextToDefaultRegionCode(number, normalizerPhoneNumber.getDialableNumber()); } }
src/main/java/de/telekom/phonenumbernormalizer/PhoneNumberNormalizerImpl.java
telekom-phonenumber-normalizer-9a1b7ad
[ { "filename": "src/main/java/de/telekom/phonenumbernormalizer/PhoneNumberNormalizer.java", "retrieved_chunk": " */\n String normalizePhoneNumber(String number, DeviceContext deviceContext);\n /**\n * Normalizes the number using PhoneLib with some additions to compensate.\n * <p>\n * Not as powerful as {@link PhoneNumberNormalizer#normalizePhoneNumber(String, DeviceContext)}, because no default NDC can be set.\n * </p>\n * @param number plain number to normalize\n * @param regionCode ISO2 code of the country, which number-plan is used for normalization\n * @return E164 formatted phone number or at least a dialable version of the number", "score": 0.9078271389007568 }, { "filename": "src/main/java/de/telekom/phonenumbernormalizer/PhoneNumberNormalizer.java", "retrieved_chunk": " /**\n * Normalizes the number using PhoneLib with some additions to compensate.\n * <p>\n * Preferable to {@link PhoneNumberNormalizer#normalizePhoneNumber(String, String)}, because default NDC can be provided, so that more compensation for generating a valid E164 can be done.\n * </p>\n * @param number plain number to normalize\n * @param deviceContext information like CC, NDC and {@link de.telekom.phonenumbernormalizer.dto.DeviceContextLineType} from which the number is dialled\n * @return E164 formatted phone number or at least a dialable version of the number\n *\n * @see PhoneNumberNormalizer#setFallbackRegionCode(String)", "score": 0.893976092338562 }, { "filename": "src/main/java/de/telekom/phonenumbernormalizer/numberplans/PhoneLibWrapper.java", "retrieved_chunk": " } else {\n if (!isSpecialFormat(dialableNumber)) {\n // Number needs normalization:\n // international prefix is added by the lib even if it's not valid in the number plan.\n this.isNormalizingTried = true;\n this.semiNormalizedNumber = PhoneLibWrapper.parseNumber(dialableNumber, regionCode);\n }\n }\n }\n }", "score": 0.8806840777397156 }, { "filename": "src/main/java/de/telekom/phonenumbernormalizer/PhoneNumberNormalizer.java", "retrieved_chunk": " */\n String normalizePhoneNumber(String number, String regionCode);\n}", "score": 0.8667299747467041 }, { "filename": "src/main/java/de/telekom/phonenumbernormalizer/numberplans/PhoneLibWrapper.java", "retrieved_chunk": " String dialableNumber;\n /**\n * The given number normalized with PhoneLib, risking we get a incorrect normalization\n *\n * @see PhoneLibWrapper#PhoneLibWrapper(String, String)\n * @see PhoneLibWrapper#isNormalizingTried()\n * @see PhoneLibWrapper#getSemiNormalizedNumber()\n */\n Phonenumber.PhoneNumber semiNormalizedNumber;\n /**", "score": 0.8666051626205444 } ]
java
(!normalizedDeviceContext.getCountryCode().equals(DeviceContext.UNKNOWN_VALUE)) {
/* * Copyright © 2023 Deutsche Telekom AG (opensource@telekom.de) * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ package de.telekom.phonenumbernormalizer; import de.telekom.phonenumbernormalizer.dto.DeviceContext; import de.telekom.phonenumbernormalizer.dto.DeviceContextMapper; import de.telekom.phonenumbernormalizer.numberplans.PhoneLibWrapper; import de.telekom.phonenumbernormalizer.numberplans.NumberPlan; import de.telekom.phonenumbernormalizer.numberplans.NumberPlanFactory; import lombok.RequiredArgsConstructor; import org.slf4j.Logger; import org.slf4j.LoggerFactory; import org.springframework.stereotype.Component; /** * Concrete implementation of {@link PhoneNumberNormalizer} using {@link PhoneLibWrapper} to normalize a number by mitigating some inaccuracies when it comes to number plans of optional NDC and NAC as zero. * <p> * Also supports {@link DeviceContext} to enrich a phone number during normalization if the optional NDC is missing. * </p> */ @RequiredArgsConstructor @Component public class PhoneNumberNormalizerImpl implements PhoneNumberNormalizer { private static final Logger LOGGER = LoggerFactory.getLogger(PhoneNumberNormalizerImpl.class); /** * Storage for {@link PhoneNumberNormalizer#setFallbackRegionCode(String)} */ private String fallbackRegionCode = null; @Override public void setFallbackRegionCode(String fallBackRegionCode) { if (PhoneLibWrapper.getCountryCodeForRegion(fallBackRegionCode) > 0) { this.fallbackRegionCode = fallBackRegionCode; } else { this.fallbackRegionCode = null; //invalid region code! } } /** * Fallback normalization within the number-plan of the fallback region. * @param number the original number to be normalized * @param dialableNumber the original number reduced to dialable digits * @return E164 formatted phone number or at least a dialable version of the number * * @see PhoneNumberNormalizer#setFallbackRegionCode(String) * @see PhoneNumberNormalizer#normalizePhoneNumber(String, String) */ private String fallbackNormalizationFromDeviceContextToDefaultRegionCode(String number, String dialableNumber) { if (this.fallbackRegionCode == null) { LOGGER.debug("Fallback Region was set!"); return dialableNumber; } else { return this.normalizePhoneNumber(number, this.fallbackRegionCode); } } /** * Uses wrapper of PhoneLib to identify if special rules apply for normalization.<br/> * Using device context for enriching the number make it normalizable to E164 format if NDC is optional in the used number plan, but not used in the phone number to be normalized. * @param wrapper instanced wrapper of PhoneLib * @param deviceContext information like CC, NDC and {@link de.telekom.phonenumbernormalizer.dto.DeviceContextLineType} from which the number is dialled * @return E164 formatted phone number or dialable version of it or null */ private String normalize(PhoneLibWrapper wrapper, DeviceContext deviceContext) { // international prefix has been added by PhoneLib even if it's not valid in the number plan. if (wrapper == null) { LOGGER.debug("PhoneLipWrapper was not initialized"); return null; } if (wrapper.getSemiNormalizedNumber() == null) { return wrapper.getDialableNumber(); } NumberPlan numberplan = null; if (deviceContext != null) { numberplan = NumberPlanFactory.INSTANCE.getNumberPlan(deviceContext.getLineType(), deviceContext.getCountryCode()); } if (wrapper.isShortNumber(numberplan)) { //if it is a short number, we can't add area code nor country code, so returning the dialable. return wrapper.getDialableNumber(); } if (wrapper.hasRegionNationalAccessCode() && deviceContext != null) { //Number plan is using a NationalPrefix aka Trunc Code ... so we could add Area Code if not included in the number. return wrapper.extendNumberByDefaultAreaCodeAndCountryCode(wrapper.getNationalAccessCode(), deviceContext.getNationalDestinationCode()); } // Number plan is not using NationalPrefix aka Trunc Code ... its also not a short number, so country code can be added: return wrapper.getE164Formatted(); } @Override public String normalizePhoneNumber(String number, String regionCode) { PhoneLibWrapper wrapper = new PhoneLibWrapper(number, regionCode); if (wrapper.getSemiNormalizedNumber() == null) { return wrapper.getDialableNumber(); } if (wrapper.isShortNumber()) { //if it is a short number, we can't add area code nor country code, so returning the dialable. return wrapper.getDialableNumber(); } // international prefix is added by the lib even if it's not valid in the number plan. //checking if the input number is equal to the nationalNumber based on number plan and trunc code logic. boolean hasNoCCAndNoNAC = wrapper.hasNoCountryCodeNorNationalAccessCode(); LOGGER.debug("Number has no CC and no NAC: {}.", hasNoCCAndNoNAC); //if the number is definitely a short number or needs an area code but does not have it, we do not add the country code. return (hasNoCCAndNoNAC) ? wrapper.getDialableNumber() : wrapper.getE164Formatted(); } @Override public String normalizePhoneNumber(String number, DeviceContext deviceContext) { // checking if the number has a special format or is not valid at all. PhoneLibWrapper normalizerPhoneNumber = new PhoneLibWrapper(number, null); if (! normalizerPhoneNumber.isNormalizingTried()) { return normalizerPhoneNumber.getDialableNumber(); } DeviceContext normalizedDeviceContext = DeviceContextMapper.normalized(deviceContext); if (!normalizedDeviceContext.getCountryCode().equals(DeviceContext.UNKNOWN_VALUE)) { String regionCode
= PhoneLibWrapper.getRegionCodeForCountryCode(normalizedDeviceContext.getCountryCode());
// now working again with the region code normalizerPhoneNumber = new PhoneLibWrapper(number, regionCode); if (!normalizedDeviceContext.getNationalDestinationCode().equals(DeviceContext.UNKNOWN_VALUE)) { // Number needs normalization: return normalize(normalizerPhoneNumber, deviceContext); } // Device Context with CountryCode but without AreaCode ... if (!(PhoneLibWrapper.UNKNOWN_REGIONCODE.equals(regionCode))) { return this.normalizePhoneNumber(number, regionCode); } } LOGGER.debug("Normalization based on DeviceContext did not work - falling back to normalization with fallback region."); return this.fallbackNormalizationFromDeviceContextToDefaultRegionCode(number, normalizerPhoneNumber.getDialableNumber()); } }
src/main/java/de/telekom/phonenumbernormalizer/PhoneNumberNormalizerImpl.java
telekom-phonenumber-normalizer-9a1b7ad
[ { "filename": "src/main/java/de/telekom/phonenumbernormalizer/PhoneNumberNormalizer.java", "retrieved_chunk": " */\n String normalizePhoneNumber(String number, DeviceContext deviceContext);\n /**\n * Normalizes the number using PhoneLib with some additions to compensate.\n * <p>\n * Not as powerful as {@link PhoneNumberNormalizer#normalizePhoneNumber(String, DeviceContext)}, because no default NDC can be set.\n * </p>\n * @param number plain number to normalize\n * @param regionCode ISO2 code of the country, which number-plan is used for normalization\n * @return E164 formatted phone number or at least a dialable version of the number", "score": 0.9198697805404663 }, { "filename": "src/main/java/de/telekom/phonenumbernormalizer/PhoneNumberNormalizer.java", "retrieved_chunk": " /**\n * Normalizes the number using PhoneLib with some additions to compensate.\n * <p>\n * Preferable to {@link PhoneNumberNormalizer#normalizePhoneNumber(String, String)}, because default NDC can be provided, so that more compensation for generating a valid E164 can be done.\n * </p>\n * @param number plain number to normalize\n * @param deviceContext information like CC, NDC and {@link de.telekom.phonenumbernormalizer.dto.DeviceContextLineType} from which the number is dialled\n * @return E164 formatted phone number or at least a dialable version of the number\n *\n * @see PhoneNumberNormalizer#setFallbackRegionCode(String)", "score": 0.8997797966003418 }, { "filename": "src/main/java/de/telekom/phonenumbernormalizer/numberplans/PhoneLibWrapper.java", "retrieved_chunk": " } else {\n if (!isSpecialFormat(dialableNumber)) {\n // Number needs normalization:\n // international prefix is added by the lib even if it's not valid in the number plan.\n this.isNormalizingTried = true;\n this.semiNormalizedNumber = PhoneLibWrapper.parseNumber(dialableNumber, regionCode);\n }\n }\n }\n }", "score": 0.890210747718811 }, { "filename": "src/main/java/de/telekom/phonenumbernormalizer/PhoneNumberNormalizer.java", "retrieved_chunk": " */\n String normalizePhoneNumber(String number, String regionCode);\n}", "score": 0.8826519250869751 }, { "filename": "src/main/java/de/telekom/phonenumbernormalizer/numberplans/PhoneLibWrapper.java", "retrieved_chunk": " String dialableNumber;\n /**\n * The given number normalized with PhoneLib, risking we get a incorrect normalization\n *\n * @see PhoneLibWrapper#PhoneLibWrapper(String, String)\n * @see PhoneLibWrapper#isNormalizingTried()\n * @see PhoneLibWrapper#getSemiNormalizedNumber()\n */\n Phonenumber.PhoneNumber semiNormalizedNumber;\n /**", "score": 0.867612361907959 } ]
java
= PhoneLibWrapper.getRegionCodeForCountryCode(normalizedDeviceContext.getCountryCode());
/* * Copyright © 2023 Deutsche Telekom AG (opensource@telekom.de) * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ package de.telekom.phonenumbernormalizer; import de.telekom.phonenumbernormalizer.dto.DeviceContext; import de.telekom.phonenumbernormalizer.dto.DeviceContextMapper; import de.telekom.phonenumbernormalizer.numberplans.PhoneLibWrapper; import de.telekom.phonenumbernormalizer.numberplans.NumberPlan; import de.telekom.phonenumbernormalizer.numberplans.NumberPlanFactory; import lombok.RequiredArgsConstructor; import org.slf4j.Logger; import org.slf4j.LoggerFactory; import org.springframework.stereotype.Component; /** * Concrete implementation of {@link PhoneNumberNormalizer} using {@link PhoneLibWrapper} to normalize a number by mitigating some inaccuracies when it comes to number plans of optional NDC and NAC as zero. * <p> * Also supports {@link DeviceContext} to enrich a phone number during normalization if the optional NDC is missing. * </p> */ @RequiredArgsConstructor @Component public class PhoneNumberNormalizerImpl implements PhoneNumberNormalizer { private static final Logger LOGGER = LoggerFactory.getLogger(PhoneNumberNormalizerImpl.class); /** * Storage for {@link PhoneNumberNormalizer#setFallbackRegionCode(String)} */ private String fallbackRegionCode = null; @Override public void setFallbackRegionCode(String fallBackRegionCode) { if (PhoneLibWrapper.getCountryCodeForRegion(fallBackRegionCode) > 0) { this.fallbackRegionCode = fallBackRegionCode; } else { this.fallbackRegionCode = null; //invalid region code! } } /** * Fallback normalization within the number-plan of the fallback region. * @param number the original number to be normalized * @param dialableNumber the original number reduced to dialable digits * @return E164 formatted phone number or at least a dialable version of the number * * @see PhoneNumberNormalizer#setFallbackRegionCode(String) * @see PhoneNumberNormalizer#normalizePhoneNumber(String, String) */ private String fallbackNormalizationFromDeviceContextToDefaultRegionCode(String number, String dialableNumber) { if (this.fallbackRegionCode == null) { LOGGER.debug("Fallback Region was set!"); return dialableNumber; } else { return this.normalizePhoneNumber(number, this.fallbackRegionCode); } } /** * Uses wrapper of PhoneLib to identify if special rules apply for normalization.<br/> * Using device context for enriching the number make it normalizable to E164 format if NDC is optional in the used number plan, but not used in the phone number to be normalized. * @param wrapper instanced wrapper of PhoneLib * @param deviceContext information like CC, NDC and {@link de.telekom.phonenumbernormalizer.dto.DeviceContextLineType} from which the number is dialled * @return E164 formatted phone number or dialable version of it or null */ private String normalize(PhoneLibWrapper wrapper, DeviceContext deviceContext) { // international prefix has been added by PhoneLib even if it's not valid in the number plan. if (wrapper == null) { LOGGER.debug("PhoneLipWrapper was not initialized"); return null; } if (wrapper.getSemiNormalizedNumber() == null) { return wrapper.getDialableNumber(); } NumberPlan numberplan = null; if (deviceContext != null) { numberplan = NumberPlanFactory.INSTANCE.getNumberPlan(deviceContext.getLineType(), deviceContext.getCountryCode()); } if (wrapper.isShortNumber(numberplan)) { //if it is a short number, we can't add area code nor country code, so returning the dialable. return wrapper.getDialableNumber(); } if (wrapper.hasRegionNationalAccessCode() && deviceContext != null) { //Number plan is using a NationalPrefix aka Trunc Code ... so we could add Area Code if not included in the number. return wrapper.extendNumberByDefaultAreaCodeAndCountryCode(wrapper.getNationalAccessCode(), deviceContext.getNationalDestinationCode()); } // Number plan is not using NationalPrefix aka Trunc Code ... its also not a short number, so country code can be added: return wrapper.getE164Formatted(); } @Override public String normalizePhoneNumber(String number, String regionCode) { PhoneLibWrapper wrapper = new PhoneLibWrapper(number, regionCode); if (wrapper.getSemiNormalizedNumber() == null) { return wrapper.getDialableNumber(); } if (wrapper.isShortNumber()) { //if it is a short number, we can't add area code nor country code, so returning the dialable. return wrapper.getDialableNumber(); } // international prefix is added by the lib even if it's not valid in the number plan. //checking if the input number is equal to the nationalNumber based on number plan and trunc code logic. boolean hasNoCCAndNoNAC = wrapper.hasNoCountryCodeNorNationalAccessCode(); LOGGER.debug("Number has no CC and no NAC: {}.", hasNoCCAndNoNAC); //if the number is definitely a short number or needs an area code but does not have it, we do not add the country code. return (hasNoCCAndNoNAC) ? wrapper.getDialableNumber() : wrapper.getE164Formatted(); } @Override public String normalizePhoneNumber(String number, DeviceContext deviceContext) { // checking if the number has a special format or is not valid at all. PhoneLibWrapper normalizerPhoneNumber = new PhoneLibWrapper(number, null);
if (! normalizerPhoneNumber.isNormalizingTried()) {
return normalizerPhoneNumber.getDialableNumber(); } DeviceContext normalizedDeviceContext = DeviceContextMapper.normalized(deviceContext); if (!normalizedDeviceContext.getCountryCode().equals(DeviceContext.UNKNOWN_VALUE)) { String regionCode = PhoneLibWrapper.getRegionCodeForCountryCode(normalizedDeviceContext.getCountryCode()); // now working again with the region code normalizerPhoneNumber = new PhoneLibWrapper(number, regionCode); if (!normalizedDeviceContext.getNationalDestinationCode().equals(DeviceContext.UNKNOWN_VALUE)) { // Number needs normalization: return normalize(normalizerPhoneNumber, deviceContext); } // Device Context with CountryCode but without AreaCode ... if (!(PhoneLibWrapper.UNKNOWN_REGIONCODE.equals(regionCode))) { return this.normalizePhoneNumber(number, regionCode); } } LOGGER.debug("Normalization based on DeviceContext did not work - falling back to normalization with fallback region."); return this.fallbackNormalizationFromDeviceContextToDefaultRegionCode(number, normalizerPhoneNumber.getDialableNumber()); } }
src/main/java/de/telekom/phonenumbernormalizer/PhoneNumberNormalizerImpl.java
telekom-phonenumber-normalizer-9a1b7ad
[ { "filename": "src/main/java/de/telekom/phonenumbernormalizer/numberplans/PhoneLibWrapper.java", "retrieved_chunk": " } else {\n if (!isSpecialFormat(dialableNumber)) {\n // Number needs normalization:\n // international prefix is added by the lib even if it's not valid in the number plan.\n this.isNormalizingTried = true;\n this.semiNormalizedNumber = PhoneLibWrapper.parseNumber(dialableNumber, regionCode);\n }\n }\n }\n }", "score": 0.8827502131462097 }, { "filename": "src/main/java/de/telekom/phonenumbernormalizer/PhoneNumberNormalizer.java", "retrieved_chunk": " */\n String normalizePhoneNumber(String number, DeviceContext deviceContext);\n /**\n * Normalizes the number using PhoneLib with some additions to compensate.\n * <p>\n * Not as powerful as {@link PhoneNumberNormalizer#normalizePhoneNumber(String, DeviceContext)}, because no default NDC can be set.\n * </p>\n * @param number plain number to normalize\n * @param regionCode ISO2 code of the country, which number-plan is used for normalization\n * @return E164 formatted phone number or at least a dialable version of the number", "score": 0.8796343803405762 }, { "filename": "src/main/java/de/telekom/phonenumbernormalizer/PhoneNumberNormalizer.java", "retrieved_chunk": " /**\n * Normalizes the number using PhoneLib with some additions to compensate.\n * <p>\n * Preferable to {@link PhoneNumberNormalizer#normalizePhoneNumber(String, String)}, because default NDC can be provided, so that more compensation for generating a valid E164 can be done.\n * </p>\n * @param number plain number to normalize\n * @param deviceContext information like CC, NDC and {@link de.telekom.phonenumbernormalizer.dto.DeviceContextLineType} from which the number is dialled\n * @return E164 formatted phone number or at least a dialable version of the number\n *\n * @see PhoneNumberNormalizer#setFallbackRegionCode(String)", "score": 0.8733383417129517 }, { "filename": "src/main/java/de/telekom/phonenumbernormalizer/numberplans/PhoneLibWrapper.java", "retrieved_chunk": " try {\n this.semiNormalizedNumber = phoneUtil.parse(extendedNumber, regionCode);\n // after area code has been added, we can add the country code by the lib:\n return getE164Formatted();\n } catch (NumberParseException e) {\n LOGGER.warn(\"could not parse extended number: {}\", extendedNumber);\n LOGGER.debug(\"{}\", e.getMessage());\n return dialableNumber;\n }\n } else {", "score": 0.8684165477752686 }, { "filename": "src/main/java/de/telekom/phonenumbernormalizer/numberplans/PhoneLibWrapper.java", "retrieved_chunk": " String dialableNumber;\n /**\n * The given number normalized with PhoneLib, risking we get a incorrect normalization\n *\n * @see PhoneLibWrapper#PhoneLibWrapper(String, String)\n * @see PhoneLibWrapper#isNormalizingTried()\n * @see PhoneLibWrapper#getSemiNormalizedNumber()\n */\n Phonenumber.PhoneNumber semiNormalizedNumber;\n /**", "score": 0.8602389097213745 } ]
java
if (! normalizerPhoneNumber.isNormalizingTried()) {
/* * Copyright © 2023 Deutsche Telekom AG (opensource@telekom.de) * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ package de.telekom.phonenumbernormalizer; import de.telekom.phonenumbernormalizer.dto.DeviceContext; import de.telekom.phonenumbernormalizer.dto.DeviceContextMapper; import de.telekom.phonenumbernormalizer.numberplans.PhoneLibWrapper; import de.telekom.phonenumbernormalizer.numberplans.NumberPlan; import de.telekom.phonenumbernormalizer.numberplans.NumberPlanFactory; import lombok.RequiredArgsConstructor; import org.slf4j.Logger; import org.slf4j.LoggerFactory; import org.springframework.stereotype.Component; /** * Concrete implementation of {@link PhoneNumberNormalizer} using {@link PhoneLibWrapper} to normalize a number by mitigating some inaccuracies when it comes to number plans of optional NDC and NAC as zero. * <p> * Also supports {@link DeviceContext} to enrich a phone number during normalization if the optional NDC is missing. * </p> */ @RequiredArgsConstructor @Component public class PhoneNumberNormalizerImpl implements PhoneNumberNormalizer { private static final Logger LOGGER = LoggerFactory.getLogger(PhoneNumberNormalizerImpl.class); /** * Storage for {@link PhoneNumberNormalizer#setFallbackRegionCode(String)} */ private String fallbackRegionCode = null; @Override public void setFallbackRegionCode(String fallBackRegionCode) { if
(PhoneLibWrapper.getCountryCodeForRegion(fallBackRegionCode) > 0) {
this.fallbackRegionCode = fallBackRegionCode; } else { this.fallbackRegionCode = null; //invalid region code! } } /** * Fallback normalization within the number-plan of the fallback region. * @param number the original number to be normalized * @param dialableNumber the original number reduced to dialable digits * @return E164 formatted phone number or at least a dialable version of the number * * @see PhoneNumberNormalizer#setFallbackRegionCode(String) * @see PhoneNumberNormalizer#normalizePhoneNumber(String, String) */ private String fallbackNormalizationFromDeviceContextToDefaultRegionCode(String number, String dialableNumber) { if (this.fallbackRegionCode == null) { LOGGER.debug("Fallback Region was set!"); return dialableNumber; } else { return this.normalizePhoneNumber(number, this.fallbackRegionCode); } } /** * Uses wrapper of PhoneLib to identify if special rules apply for normalization.<br/> * Using device context for enriching the number make it normalizable to E164 format if NDC is optional in the used number plan, but not used in the phone number to be normalized. * @param wrapper instanced wrapper of PhoneLib * @param deviceContext information like CC, NDC and {@link de.telekom.phonenumbernormalizer.dto.DeviceContextLineType} from which the number is dialled * @return E164 formatted phone number or dialable version of it or null */ private String normalize(PhoneLibWrapper wrapper, DeviceContext deviceContext) { // international prefix has been added by PhoneLib even if it's not valid in the number plan. if (wrapper == null) { LOGGER.debug("PhoneLipWrapper was not initialized"); return null; } if (wrapper.getSemiNormalizedNumber() == null) { return wrapper.getDialableNumber(); } NumberPlan numberplan = null; if (deviceContext != null) { numberplan = NumberPlanFactory.INSTANCE.getNumberPlan(deviceContext.getLineType(), deviceContext.getCountryCode()); } if (wrapper.isShortNumber(numberplan)) { //if it is a short number, we can't add area code nor country code, so returning the dialable. return wrapper.getDialableNumber(); } if (wrapper.hasRegionNationalAccessCode() && deviceContext != null) { //Number plan is using a NationalPrefix aka Trunc Code ... so we could add Area Code if not included in the number. return wrapper.extendNumberByDefaultAreaCodeAndCountryCode(wrapper.getNationalAccessCode(), deviceContext.getNationalDestinationCode()); } // Number plan is not using NationalPrefix aka Trunc Code ... its also not a short number, so country code can be added: return wrapper.getE164Formatted(); } @Override public String normalizePhoneNumber(String number, String regionCode) { PhoneLibWrapper wrapper = new PhoneLibWrapper(number, regionCode); if (wrapper.getSemiNormalizedNumber() == null) { return wrapper.getDialableNumber(); } if (wrapper.isShortNumber()) { //if it is a short number, we can't add area code nor country code, so returning the dialable. return wrapper.getDialableNumber(); } // international prefix is added by the lib even if it's not valid in the number plan. //checking if the input number is equal to the nationalNumber based on number plan and trunc code logic. boolean hasNoCCAndNoNAC = wrapper.hasNoCountryCodeNorNationalAccessCode(); LOGGER.debug("Number has no CC and no NAC: {}.", hasNoCCAndNoNAC); //if the number is definitely a short number or needs an area code but does not have it, we do not add the country code. return (hasNoCCAndNoNAC) ? wrapper.getDialableNumber() : wrapper.getE164Formatted(); } @Override public String normalizePhoneNumber(String number, DeviceContext deviceContext) { // checking if the number has a special format or is not valid at all. PhoneLibWrapper normalizerPhoneNumber = new PhoneLibWrapper(number, null); if (! normalizerPhoneNumber.isNormalizingTried()) { return normalizerPhoneNumber.getDialableNumber(); } DeviceContext normalizedDeviceContext = DeviceContextMapper.normalized(deviceContext); if (!normalizedDeviceContext.getCountryCode().equals(DeviceContext.UNKNOWN_VALUE)) { String regionCode = PhoneLibWrapper.getRegionCodeForCountryCode(normalizedDeviceContext.getCountryCode()); // now working again with the region code normalizerPhoneNumber = new PhoneLibWrapper(number, regionCode); if (!normalizedDeviceContext.getNationalDestinationCode().equals(DeviceContext.UNKNOWN_VALUE)) { // Number needs normalization: return normalize(normalizerPhoneNumber, deviceContext); } // Device Context with CountryCode but without AreaCode ... if (!(PhoneLibWrapper.UNKNOWN_REGIONCODE.equals(regionCode))) { return this.normalizePhoneNumber(number, regionCode); } } LOGGER.debug("Normalization based on DeviceContext did not work - falling back to normalization with fallback region."); return this.fallbackNormalizationFromDeviceContextToDefaultRegionCode(number, normalizerPhoneNumber.getDialableNumber()); } }
src/main/java/de/telekom/phonenumbernormalizer/PhoneNumberNormalizerImpl.java
telekom-phonenumber-normalizer-9a1b7ad
[ { "filename": "src/main/java/de/telekom/phonenumbernormalizer/PhoneNumberNormalizer.java", "retrieved_chunk": "public interface PhoneNumberNormalizer {\n /**\n * Sets the ISO2 country code, which is used if the {@link DeviceContext} does not indicate one.\n * The country might represent a region, like \"US\" for North America.\n *\n * @param fallBackRegionCode ISO2 code of the country\n *\n * @see PhoneNumberNormalizer#normalizePhoneNumber(String, DeviceContext)\n */\n void setFallbackRegionCode(String fallBackRegionCode);", "score": 0.9129412174224854 }, { "filename": "src/main/java/de/telekom/phonenumbernormalizer/numberplans/PhoneLibWrapper.java", "retrieved_chunk": " } else {\n if (!isSpecialFormat(dialableNumber)) {\n // Number needs normalization:\n // international prefix is added by the lib even if it's not valid in the number plan.\n this.isNormalizingTried = true;\n this.semiNormalizedNumber = PhoneLibWrapper.parseNumber(dialableNumber, regionCode);\n }\n }\n }\n }", "score": 0.8377540111541748 }, { "filename": "src/main/java/de/telekom/phonenumbernormalizer/PhoneNumberNormalizer.java", "retrieved_chunk": " */\n String normalizePhoneNumber(String number, String regionCode);\n}", "score": 0.8373737335205078 }, { "filename": "src/main/java/de/telekom/phonenumbernormalizer/numberplans/PhoneLibWrapper.java", "retrieved_chunk": " * @param number the phone number to be analyzed\n * @param regionCode the ISO2 Code of the Region / Country, which telephone number plan is used\n */\n public PhoneLibWrapper(String number, String regionCode) {\n this.regionCode = regionCode;\n this.metadata = getMetadataForRegion();\n if (number != null) {\n this.dialableNumber = PhoneNumberUtil.normalizeDiallableCharsOnly(number);\n if (this.dialableNumber.isEmpty()) {\n this.dialableNumber = \"\";", "score": 0.8357512950897217 }, { "filename": "src/main/java/de/telekom/phonenumbernormalizer/numberplans/PhoneLibWrapper.java", "retrieved_chunk": " * @param regionCode ISO2 code for the regions number plan used for parsing the number\n * @return either the parsed {@link Phonenumber.PhoneNumber} or null\n */\n private static Phonenumber.PhoneNumber parseNumber(String number, String regionCode) {\n try {\n return phoneUtil.parse(number, regionCode);\n // international prefix is added by the lib even if it's not valid in the number plan.\n } catch (NumberParseException e) {\n LOGGER.warn(\"could not parse normalize number: {}\", number);\n LOGGER.debug(\"{}\", e.getMessage());", "score": 0.8296087980270386 } ]
java
(PhoneLibWrapper.getCountryCodeForRegion(fallBackRegionCode) > 0) {
/* * Copyright © 2023 Deutsche Telekom AG (opensource@telekom.de) * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ package de.telekom.phonenumbernormalizer; import de.telekom.phonenumbernormalizer.dto.DeviceContext; import de.telekom.phonenumbernormalizer.dto.DeviceContextMapper; import de.telekom.phonenumbernormalizer.numberplans.PhoneLibWrapper; import de.telekom.phonenumbernormalizer.numberplans.NumberPlan; import de.telekom.phonenumbernormalizer.numberplans.NumberPlanFactory; import lombok.RequiredArgsConstructor; import org.slf4j.Logger; import org.slf4j.LoggerFactory; import org.springframework.stereotype.Component; /** * Concrete implementation of {@link PhoneNumberNormalizer} using {@link PhoneLibWrapper} to normalize a number by mitigating some inaccuracies when it comes to number plans of optional NDC and NAC as zero. * <p> * Also supports {@link DeviceContext} to enrich a phone number during normalization if the optional NDC is missing. * </p> */ @RequiredArgsConstructor @Component public class PhoneNumberNormalizerImpl implements PhoneNumberNormalizer { private static final Logger LOGGER = LoggerFactory.getLogger(PhoneNumberNormalizerImpl.class); /** * Storage for {@link PhoneNumberNormalizer#setFallbackRegionCode(String)} */ private String fallbackRegionCode = null; @Override public void setFallbackRegionCode(String fallBackRegionCode) { if (PhoneLibWrapper.getCountryCodeForRegion(fallBackRegionCode) > 0) { this.fallbackRegionCode = fallBackRegionCode; } else { this.fallbackRegionCode = null; //invalid region code! } } /** * Fallback normalization within the number-plan of the fallback region. * @param number the original number to be normalized * @param dialableNumber the original number reduced to dialable digits * @return E164 formatted phone number or at least a dialable version of the number * * @see PhoneNumberNormalizer#setFallbackRegionCode(String) * @see PhoneNumberNormalizer#normalizePhoneNumber(String, String) */ private String fallbackNormalizationFromDeviceContextToDefaultRegionCode(String number, String dialableNumber) { if (this.fallbackRegionCode == null) { LOGGER.debug("Fallback Region was set!"); return dialableNumber; } else { return this.normalizePhoneNumber(number, this.fallbackRegionCode); } } /** * Uses wrapper of PhoneLib to identify if special rules apply for normalization.<br/> * Using device context for enriching the number make it normalizable to E164 format if NDC is optional in the used number plan, but not used in the phone number to be normalized. * @param wrapper instanced wrapper of PhoneLib * @param deviceContext information like CC, NDC and {@link de.telekom.phonenumbernormalizer.dto.DeviceContextLineType} from which the number is dialled * @return E164 formatted phone number or dialable version of it or null */ private String normalize(PhoneLibWrapper wrapper, DeviceContext deviceContext) { // international prefix has been added by PhoneLib even if it's not valid in the number plan. if (wrapper == null) { LOGGER.debug("PhoneLipWrapper was not initialized"); return null; } if (wrapper.getSemiNormalizedNumber() == null) { return wrapper.getDialableNumber(); } NumberPlan numberplan = null; if (deviceContext != null) { numberplan = NumberPlanFactory.INSTANCE.getNumberPlan(deviceContext.getLineType(), deviceContext.getCountryCode()); } if (wrapper.isShortNumber(numberplan)) { //if it is a short number, we can't add area code nor country code, so returning the dialable. return wrapper.getDialableNumber(); } if
(wrapper.hasRegionNationalAccessCode() && deviceContext != null) {
//Number plan is using a NationalPrefix aka Trunc Code ... so we could add Area Code if not included in the number. return wrapper.extendNumberByDefaultAreaCodeAndCountryCode(wrapper.getNationalAccessCode(), deviceContext.getNationalDestinationCode()); } // Number plan is not using NationalPrefix aka Trunc Code ... its also not a short number, so country code can be added: return wrapper.getE164Formatted(); } @Override public String normalizePhoneNumber(String number, String regionCode) { PhoneLibWrapper wrapper = new PhoneLibWrapper(number, regionCode); if (wrapper.getSemiNormalizedNumber() == null) { return wrapper.getDialableNumber(); } if (wrapper.isShortNumber()) { //if it is a short number, we can't add area code nor country code, so returning the dialable. return wrapper.getDialableNumber(); } // international prefix is added by the lib even if it's not valid in the number plan. //checking if the input number is equal to the nationalNumber based on number plan and trunc code logic. boolean hasNoCCAndNoNAC = wrapper.hasNoCountryCodeNorNationalAccessCode(); LOGGER.debug("Number has no CC and no NAC: {}.", hasNoCCAndNoNAC); //if the number is definitely a short number or needs an area code but does not have it, we do not add the country code. return (hasNoCCAndNoNAC) ? wrapper.getDialableNumber() : wrapper.getE164Formatted(); } @Override public String normalizePhoneNumber(String number, DeviceContext deviceContext) { // checking if the number has a special format or is not valid at all. PhoneLibWrapper normalizerPhoneNumber = new PhoneLibWrapper(number, null); if (! normalizerPhoneNumber.isNormalizingTried()) { return normalizerPhoneNumber.getDialableNumber(); } DeviceContext normalizedDeviceContext = DeviceContextMapper.normalized(deviceContext); if (!normalizedDeviceContext.getCountryCode().equals(DeviceContext.UNKNOWN_VALUE)) { String regionCode = PhoneLibWrapper.getRegionCodeForCountryCode(normalizedDeviceContext.getCountryCode()); // now working again with the region code normalizerPhoneNumber = new PhoneLibWrapper(number, regionCode); if (!normalizedDeviceContext.getNationalDestinationCode().equals(DeviceContext.UNKNOWN_VALUE)) { // Number needs normalization: return normalize(normalizerPhoneNumber, deviceContext); } // Device Context with CountryCode but without AreaCode ... if (!(PhoneLibWrapper.UNKNOWN_REGIONCODE.equals(regionCode))) { return this.normalizePhoneNumber(number, regionCode); } } LOGGER.debug("Normalization based on DeviceContext did not work - falling back to normalization with fallback region."); return this.fallbackNormalizationFromDeviceContextToDefaultRegionCode(number, normalizerPhoneNumber.getDialableNumber()); } }
src/main/java/de/telekom/phonenumbernormalizer/PhoneNumberNormalizerImpl.java
telekom-phonenumber-normalizer-9a1b7ad
[ { "filename": "src/main/java/de/telekom/phonenumbernormalizer/numberplans/PhoneLibWrapper.java", "retrieved_chunk": " } else {\n if (!isSpecialFormat(dialableNumber)) {\n // Number needs normalization:\n // international prefix is added by the lib even if it's not valid in the number plan.\n this.isNormalizingTried = true;\n this.semiNormalizedNumber = PhoneLibWrapper.parseNumber(dialableNumber, regionCode);\n }\n }\n }\n }", "score": 0.877816915512085 }, { "filename": "src/main/java/de/telekom/phonenumbernormalizer/numberplans/PhoneLibWrapper.java", "retrieved_chunk": " try {\n this.semiNormalizedNumber = phoneUtil.parse(extendedNumber, regionCode);\n // after area code has been added, we can add the country code by the lib:\n return getE164Formatted();\n } catch (NumberParseException e) {\n LOGGER.warn(\"could not parse extended number: {}\", extendedNumber);\n LOGGER.debug(\"{}\", e.getMessage());\n return dialableNumber;\n }\n } else {", "score": 0.8691015839576721 }, { "filename": "src/main/java/de/telekom/phonenumbernormalizer/numberplans/PhoneLibWrapper.java", "retrieved_chunk": " * If we have a plain national number based on regions number plan and potential NAC logic.\n * <p>\n * For a number plan without NAC logic, it will always return false!\n * </p>\n * @return if given number could have CC and NAC, but does not have any of them.\n */\n public boolean hasNoCountryCodeNorNationalAccessCode() {\n // if given number has no NAC and no CC, it equals national phone number (without NAC).\n if (! Objects.equals(dialableNumber, this.getNationalPhoneNumberWithoutNationalAccessCode())) {\n return false;", "score": 0.8597544431686401 }, { "filename": "src/main/java/de/telekom/phonenumbernormalizer/numberplans/PhoneLibWrapper.java", "retrieved_chunk": " *\n * @see PhoneLibWrapper#PhoneLibWrapper(String, String)\n */\n public boolean isShortNumber(NumberPlan numberplan) {\n if (numberplan == null) {\n return this.isShortNumber();\n }\n return numberplan.isMatchingShortNumber(this.getDialableNumber());\n }\n /**", "score": 0.8585328459739685 }, { "filename": "src/main/java/de/telekom/phonenumbernormalizer/numberplans/NumberPlanFactory.java", "retrieved_chunk": " */\n public NumberPlan getNumberPlan(DeviceContextLineType numberPlanType, String countryCode) {\n if (numberPlans.containsKey(numberPlanType)) {\n LOGGER.debug(\"use number plan for type: {}\", numberPlanType);\n Map<String, NumberPlan> numberPlan = numberPlans.get(numberPlanType);\n if (numberPlan.containsKey(countryCode)) {\n LOGGER.debug(\"use number plan for country code: {}\", countryCode);\n return numberPlan.get(countryCode);\n }\n }", "score": 0.8567874431610107 } ]
java
(wrapper.hasRegionNationalAccessCode() && deviceContext != null) {
/* * Copyright © 2023 Deutsche Telekom AG (opensource@telekom.de) * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ package de.telekom.phonenumbernormalizer; import de.telekom.phonenumbernormalizer.dto.DeviceContext; import de.telekom.phonenumbernormalizer.dto.DeviceContextMapper; import de.telekom.phonenumbernormalizer.numberplans.PhoneLibWrapper; import de.telekom.phonenumbernormalizer.numberplans.NumberPlan; import de.telekom.phonenumbernormalizer.numberplans.NumberPlanFactory; import lombok.RequiredArgsConstructor; import org.slf4j.Logger; import org.slf4j.LoggerFactory; import org.springframework.stereotype.Component; /** * Concrete implementation of {@link PhoneNumberNormalizer} using {@link PhoneLibWrapper} to normalize a number by mitigating some inaccuracies when it comes to number plans of optional NDC and NAC as zero. * <p> * Also supports {@link DeviceContext} to enrich a phone number during normalization if the optional NDC is missing. * </p> */ @RequiredArgsConstructor @Component public class PhoneNumberNormalizerImpl implements PhoneNumberNormalizer { private static final Logger LOGGER = LoggerFactory.getLogger(PhoneNumberNormalizerImpl.class); /** * Storage for {@link PhoneNumberNormalizer#setFallbackRegionCode(String)} */ private String fallbackRegionCode = null; @Override public void setFallbackRegionCode(String fallBackRegionCode) { if (PhoneLibWrapper.getCountryCodeForRegion(fallBackRegionCode) > 0) { this.fallbackRegionCode = fallBackRegionCode; } else { this.fallbackRegionCode = null; //invalid region code! } } /** * Fallback normalization within the number-plan of the fallback region. * @param number the original number to be normalized * @param dialableNumber the original number reduced to dialable digits * @return E164 formatted phone number or at least a dialable version of the number * * @see PhoneNumberNormalizer#setFallbackRegionCode(String) * @see PhoneNumberNormalizer#normalizePhoneNumber(String, String) */ private String fallbackNormalizationFromDeviceContextToDefaultRegionCode(String number, String dialableNumber) { if (this.fallbackRegionCode == null) { LOGGER.debug("Fallback Region was set!"); return dialableNumber; } else { return this.normalizePhoneNumber(number, this.fallbackRegionCode); } } /** * Uses wrapper of PhoneLib to identify if special rules apply for normalization.<br/> * Using device context for enriching the number make it normalizable to E164 format if NDC is optional in the used number plan, but not used in the phone number to be normalized. * @param wrapper instanced wrapper of PhoneLib * @param deviceContext information like CC, NDC and {@link de.telekom.phonenumbernormalizer.dto.DeviceContextLineType} from which the number is dialled * @return E164 formatted phone number or dialable version of it or null */ private String normalize(PhoneLibWrapper wrapper, DeviceContext deviceContext) { // international prefix has been added by PhoneLib even if it's not valid in the number plan. if (wrapper == null) { LOGGER.debug("PhoneLipWrapper was not initialized"); return null; } if (wrapper.getSemiNormalizedNumber() == null) { return wrapper.getDialableNumber(); } NumberPlan numberplan = null; if (deviceContext != null) { numberplan = NumberPlanFactory.INSTANCE.getNumberPlan(deviceContext.getLineType(), deviceContext.getCountryCode()); } if (wrapper.isShortNumber(numberplan)) { //if it is a short number, we can't add area code nor country code, so returning the dialable. return wrapper.getDialableNumber(); } if (wrapper.hasRegionNationalAccessCode() && deviceContext != null) { //Number plan is using a NationalPrefix aka Trunc Code ... so we could add Area Code if not included in the number. return wrapper
.extendNumberByDefaultAreaCodeAndCountryCode(wrapper.getNationalAccessCode(), deviceContext.getNationalDestinationCode());
} // Number plan is not using NationalPrefix aka Trunc Code ... its also not a short number, so country code can be added: return wrapper.getE164Formatted(); } @Override public String normalizePhoneNumber(String number, String regionCode) { PhoneLibWrapper wrapper = new PhoneLibWrapper(number, regionCode); if (wrapper.getSemiNormalizedNumber() == null) { return wrapper.getDialableNumber(); } if (wrapper.isShortNumber()) { //if it is a short number, we can't add area code nor country code, so returning the dialable. return wrapper.getDialableNumber(); } // international prefix is added by the lib even if it's not valid in the number plan. //checking if the input number is equal to the nationalNumber based on number plan and trunc code logic. boolean hasNoCCAndNoNAC = wrapper.hasNoCountryCodeNorNationalAccessCode(); LOGGER.debug("Number has no CC and no NAC: {}.", hasNoCCAndNoNAC); //if the number is definitely a short number or needs an area code but does not have it, we do not add the country code. return (hasNoCCAndNoNAC) ? wrapper.getDialableNumber() : wrapper.getE164Formatted(); } @Override public String normalizePhoneNumber(String number, DeviceContext deviceContext) { // checking if the number has a special format or is not valid at all. PhoneLibWrapper normalizerPhoneNumber = new PhoneLibWrapper(number, null); if (! normalizerPhoneNumber.isNormalizingTried()) { return normalizerPhoneNumber.getDialableNumber(); } DeviceContext normalizedDeviceContext = DeviceContextMapper.normalized(deviceContext); if (!normalizedDeviceContext.getCountryCode().equals(DeviceContext.UNKNOWN_VALUE)) { String regionCode = PhoneLibWrapper.getRegionCodeForCountryCode(normalizedDeviceContext.getCountryCode()); // now working again with the region code normalizerPhoneNumber = new PhoneLibWrapper(number, regionCode); if (!normalizedDeviceContext.getNationalDestinationCode().equals(DeviceContext.UNKNOWN_VALUE)) { // Number needs normalization: return normalize(normalizerPhoneNumber, deviceContext); } // Device Context with CountryCode but without AreaCode ... if (!(PhoneLibWrapper.UNKNOWN_REGIONCODE.equals(regionCode))) { return this.normalizePhoneNumber(number, regionCode); } } LOGGER.debug("Normalization based on DeviceContext did not work - falling back to normalization with fallback region."); return this.fallbackNormalizationFromDeviceContextToDefaultRegionCode(number, normalizerPhoneNumber.getDialableNumber()); } }
src/main/java/de/telekom/phonenumbernormalizer/PhoneNumberNormalizerImpl.java
telekom-phonenumber-normalizer-9a1b7ad
[ { "filename": "src/main/java/de/telekom/phonenumbernormalizer/numberplans/PhoneLibWrapper.java", "retrieved_chunk": " } else {\n if (!isSpecialFormat(dialableNumber)) {\n // Number needs normalization:\n // international prefix is added by the lib even if it's not valid in the number plan.\n this.isNormalizingTried = true;\n this.semiNormalizedNumber = PhoneLibWrapper.parseNumber(dialableNumber, regionCode);\n }\n }\n }\n }", "score": 0.8707174062728882 }, { "filename": "src/main/java/de/telekom/phonenumbernormalizer/numberplans/PhoneLibWrapper.java", "retrieved_chunk": " try {\n this.semiNormalizedNumber = phoneUtil.parse(extendedNumber, regionCode);\n // after area code has been added, we can add the country code by the lib:\n return getE164Formatted();\n } catch (NumberParseException e) {\n LOGGER.warn(\"could not parse extended number: {}\", extendedNumber);\n LOGGER.debug(\"{}\", e.getMessage());\n return dialableNumber;\n }\n } else {", "score": 0.8667348623275757 }, { "filename": "src/main/java/de/telekom/phonenumbernormalizer/numberplans/NumberPlanFactory.java", "retrieved_chunk": " */\n public NumberPlan getNumberPlan(DeviceContextLineType numberPlanType, String countryCode) {\n if (numberPlans.containsKey(numberPlanType)) {\n LOGGER.debug(\"use number plan for type: {}\", numberPlanType);\n Map<String, NumberPlan> numberPlan = numberPlans.get(numberPlanType);\n if (numberPlan.containsKey(countryCode)) {\n LOGGER.debug(\"use number plan for country code: {}\", countryCode);\n return numberPlan.get(countryCode);\n }\n }", "score": 0.8603444695472717 }, { "filename": "src/main/java/de/telekom/phonenumbernormalizer/numberplans/PhoneLibWrapper.java", "retrieved_chunk": " * @param defaultNationalDestinationCode the NDC to be added depending on the use telephone line origination.\n * @return if possible a E164 formatted representation or just the diallable representation of the given number.\n *\n * @see PhoneLibWrapper#PhoneLibWrapper(String, String)\n */\n public String extendNumberByDefaultAreaCodeAndCountryCode(String nationalAccessCode, String defaultNationalDestinationCode) {\n String nationalPhoneNumberWithoutNationalAccessCode = this.getNationalPhoneNumberWithoutNationalAccessCode();\n //if the dialableNumber is same as the national Number, Without NationalPrefix, then there is no NDC, so it needs to be added.\n if (Objects.equals(dialableNumber, nationalPhoneNumberWithoutNationalAccessCode)) {\n String extendedNumber = nationalAccessCode + defaultNationalDestinationCode + nationalPhoneNumberWithoutNationalAccessCode;", "score": 0.8569250106811523 }, { "filename": "src/main/java/de/telekom/phonenumbernormalizer/numberplans/PhoneLibWrapper.java", "retrieved_chunk": " * If we have a plain national number based on regions number plan and potential NAC logic.\n * <p>\n * For a number plan without NAC logic, it will always return false!\n * </p>\n * @return if given number could have CC and NAC, but does not have any of them.\n */\n public boolean hasNoCountryCodeNorNationalAccessCode() {\n // if given number has no NAC and no CC, it equals national phone number (without NAC).\n if (! Objects.equals(dialableNumber, this.getNationalPhoneNumberWithoutNationalAccessCode())) {\n return false;", "score": 0.8499487638473511 } ]
java
.extendNumberByDefaultAreaCodeAndCountryCode(wrapper.getNationalAccessCode(), deviceContext.getNationalDestinationCode());
/* * Copyright © 2023 Deutsche Telekom AG (opensource@telekom.de) * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ package de.telekom.phonenumbernormalizer; import de.telekom.phonenumbernormalizer.dto.DeviceContext; import de.telekom.phonenumbernormalizer.dto.DeviceContextMapper; import de.telekom.phonenumbernormalizer.numberplans.PhoneLibWrapper; import de.telekom.phonenumbernormalizer.numberplans.NumberPlan; import de.telekom.phonenumbernormalizer.numberplans.NumberPlanFactory; import lombok.RequiredArgsConstructor; import org.slf4j.Logger; import org.slf4j.LoggerFactory; import org.springframework.stereotype.Component; /** * Concrete implementation of {@link PhoneNumberNormalizer} using {@link PhoneLibWrapper} to normalize a number by mitigating some inaccuracies when it comes to number plans of optional NDC and NAC as zero. * <p> * Also supports {@link DeviceContext} to enrich a phone number during normalization if the optional NDC is missing. * </p> */ @RequiredArgsConstructor @Component public class PhoneNumberNormalizerImpl implements PhoneNumberNormalizer { private static final Logger LOGGER = LoggerFactory.getLogger(PhoneNumberNormalizerImpl.class); /** * Storage for {@link PhoneNumberNormalizer#setFallbackRegionCode(String)} */ private String fallbackRegionCode = null; @Override public void setFallbackRegionCode(String fallBackRegionCode) { if (PhoneLibWrapper.getCountryCodeForRegion(fallBackRegionCode) > 0) { this.fallbackRegionCode = fallBackRegionCode; } else { this.fallbackRegionCode = null; //invalid region code! } } /** * Fallback normalization within the number-plan of the fallback region. * @param number the original number to be normalized * @param dialableNumber the original number reduced to dialable digits * @return E164 formatted phone number or at least a dialable version of the number * * @see PhoneNumberNormalizer#setFallbackRegionCode(String) * @see PhoneNumberNormalizer#normalizePhoneNumber(String, String) */ private String fallbackNormalizationFromDeviceContextToDefaultRegionCode(String number, String dialableNumber) { if (this.fallbackRegionCode == null) { LOGGER.debug("Fallback Region was set!"); return dialableNumber; } else { return this.normalizePhoneNumber(number, this.fallbackRegionCode); } } /** * Uses wrapper of PhoneLib to identify if special rules apply for normalization.<br/> * Using device context for enriching the number make it normalizable to E164 format if NDC is optional in the used number plan, but not used in the phone number to be normalized. * @param wrapper instanced wrapper of PhoneLib * @param deviceContext information like CC, NDC and {@link de.telekom.phonenumbernormalizer.dto.DeviceContextLineType} from which the number is dialled * @return E164 formatted phone number or dialable version of it or null */ private String normalize(PhoneLibWrapper wrapper, DeviceContext deviceContext) { // international prefix has been added by PhoneLib even if it's not valid in the number plan. if (wrapper == null) { LOGGER.debug("PhoneLipWrapper was not initialized"); return null; } if (wrapper.getSemiNormalizedNumber() == null) { return wrapper.getDialableNumber(); } NumberPlan numberplan = null; if (deviceContext != null) { numberplan = NumberPlanFactory.INSTANCE.getNumberPlan(deviceContext.getLineType(), deviceContext.getCountryCode()); } if (wrapper.isShortNumber(numberplan)) { //if it is a short number, we can't add area code nor country code, so returning the dialable. return wrapper.getDialableNumber(); } if (wrapper.hasRegionNationalAccessCode() && deviceContext != null) { //Number plan is using a NationalPrefix aka Trunc Code ... so we could add Area Code if not included in the number. return wrapper.extendNumberByDefaultAreaCodeAndCountryCode(wrapper.getNationalAccessCode(), deviceContext.getNationalDestinationCode()); } // Number plan is not using NationalPrefix aka Trunc Code ... its also not a short number, so country code can be added: return wrapper.getE164Formatted(); } @Override public String normalizePhoneNumber(String number, String regionCode) { PhoneLibWrapper wrapper = new PhoneLibWrapper(number, regionCode); if (wrapper.getSemiNormalizedNumber() == null) { return wrapper.getDialableNumber(); } if (wrapper.isShortNumber()) { //if it is a short number, we can't add area code nor country code, so returning the dialable. return wrapper.getDialableNumber(); } // international prefix is added by the lib even if it's not valid in the number plan. //checking if the input number is equal to the nationalNumber based on number plan and trunc code logic. boolean hasNoCCAndNoNAC = wrapper.hasNoCountryCodeNorNationalAccessCode(); LOGGER.debug("Number has no CC and no NAC: {}.", hasNoCCAndNoNAC); //if the number is definitely a short number or needs an area code but does not have it, we do not add the country code. return (hasNoCCAndNoNAC) ? wrapper.getDialableNumber(
) : wrapper.getE164Formatted();
} @Override public String normalizePhoneNumber(String number, DeviceContext deviceContext) { // checking if the number has a special format or is not valid at all. PhoneLibWrapper normalizerPhoneNumber = new PhoneLibWrapper(number, null); if (! normalizerPhoneNumber.isNormalizingTried()) { return normalizerPhoneNumber.getDialableNumber(); } DeviceContext normalizedDeviceContext = DeviceContextMapper.normalized(deviceContext); if (!normalizedDeviceContext.getCountryCode().equals(DeviceContext.UNKNOWN_VALUE)) { String regionCode = PhoneLibWrapper.getRegionCodeForCountryCode(normalizedDeviceContext.getCountryCode()); // now working again with the region code normalizerPhoneNumber = new PhoneLibWrapper(number, regionCode); if (!normalizedDeviceContext.getNationalDestinationCode().equals(DeviceContext.UNKNOWN_VALUE)) { // Number needs normalization: return normalize(normalizerPhoneNumber, deviceContext); } // Device Context with CountryCode but without AreaCode ... if (!(PhoneLibWrapper.UNKNOWN_REGIONCODE.equals(regionCode))) { return this.normalizePhoneNumber(number, regionCode); } } LOGGER.debug("Normalization based on DeviceContext did not work - falling back to normalization with fallback region."); return this.fallbackNormalizationFromDeviceContextToDefaultRegionCode(number, normalizerPhoneNumber.getDialableNumber()); } }
src/main/java/de/telekom/phonenumbernormalizer/PhoneNumberNormalizerImpl.java
telekom-phonenumber-normalizer-9a1b7ad
[ { "filename": "src/main/java/de/telekom/phonenumbernormalizer/numberplans/PhoneLibWrapper.java", "retrieved_chunk": " try {\n this.semiNormalizedNumber = phoneUtil.parse(extendedNumber, regionCode);\n // after area code has been added, we can add the country code by the lib:\n return getE164Formatted();\n } catch (NumberParseException e) {\n LOGGER.warn(\"could not parse extended number: {}\", extendedNumber);\n LOGGER.debug(\"{}\", e.getMessage());\n return dialableNumber;\n }\n } else {", "score": 0.8779116272926331 }, { "filename": "src/main/java/de/telekom/phonenumbernormalizer/numberplans/PhoneLibWrapper.java", "retrieved_chunk": " *\n * @return national number without NAC, but any other leading zero.\n *\n * @see PhoneLibWrapper#PhoneLibWrapper(String, String)\n * @see PhoneLibWrapper#getSemiNormalizedNumber()\n * @see PhoneLibWrapper#nationalPhoneNumberWithoutNationalPrefix(Phonenumber.PhoneNumber)\n */\n private String getNationalPhoneNumberWithoutNationalAccessCode() {\n return PhoneLibWrapper.nationalPhoneNumberWithoutNationalPrefix(this.semiNormalizedNumber);\n }", "score": 0.875759482383728 }, { "filename": "src/main/java/de/telekom/phonenumbernormalizer/numberplans/PhoneLibWrapper.java", "retrieved_chunk": " } else {\n if (!isSpecialFormat(dialableNumber)) {\n // Number needs normalization:\n // international prefix is added by the lib even if it's not valid in the number plan.\n this.isNormalizingTried = true;\n this.semiNormalizedNumber = PhoneLibWrapper.parseNumber(dialableNumber, regionCode);\n }\n }\n }\n }", "score": 0.8738796710968018 }, { "filename": "src/main/java/de/telekom/phonenumbernormalizer/numberplans/PhoneLibWrapper.java", "retrieved_chunk": " *\n * @see PhoneLibWrapper#PhoneLibWrapper(String, String)\n */\n public String getE164Formatted() {\n return phoneUtil.format(this.semiNormalizedNumber, PhoneNumberUtil.PhoneNumberFormat.E164);\n }\n /**\n * If we know the given region for the given number {@link PhoneLibWrapper#hasRegionNationalAccessCode()}, this method checks if the given number does not start with a NAC nor a CC,\n * so we could permanently add a default NDC and NAC to the given number and for this new value the method directly return a E164 formatted representation.\n * @param nationalAccessCode the NAC to be added e.g. for Germany it would be \"0\"", "score": 0.8724063634872437 }, { "filename": "src/main/java/de/telekom/phonenumbernormalizer/numberplans/PhoneLibWrapper.java", "retrieved_chunk": " * @param defaultNationalDestinationCode the NDC to be added depending on the use telephone line origination.\n * @return if possible a E164 formatted representation or just the diallable representation of the given number.\n *\n * @see PhoneLibWrapper#PhoneLibWrapper(String, String)\n */\n public String extendNumberByDefaultAreaCodeAndCountryCode(String nationalAccessCode, String defaultNationalDestinationCode) {\n String nationalPhoneNumberWithoutNationalAccessCode = this.getNationalPhoneNumberWithoutNationalAccessCode();\n //if the dialableNumber is same as the national Number, Without NationalPrefix, then there is no NDC, so it needs to be added.\n if (Objects.equals(dialableNumber, nationalPhoneNumberWithoutNationalAccessCode)) {\n String extendedNumber = nationalAccessCode + defaultNationalDestinationCode + nationalPhoneNumberWithoutNationalAccessCode;", "score": 0.8650097846984863 } ]
java
) : wrapper.getE164Formatted();
/* * Copyright © 2023 Deutsche Telekom AG (opensource@telekom.de) * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ package de.telekom.phonenumbernormalizer; import de.telekom.phonenumbernormalizer.dto.DeviceContext; import de.telekom.phonenumbernormalizer.dto.DeviceContextMapper; import de.telekom.phonenumbernormalizer.numberplans.PhoneLibWrapper; import de.telekom.phonenumbernormalizer.numberplans.NumberPlan; import de.telekom.phonenumbernormalizer.numberplans.NumberPlanFactory; import lombok.RequiredArgsConstructor; import org.slf4j.Logger; import org.slf4j.LoggerFactory; import org.springframework.stereotype.Component; /** * Concrete implementation of {@link PhoneNumberNormalizer} using {@link PhoneLibWrapper} to normalize a number by mitigating some inaccuracies when it comes to number plans of optional NDC and NAC as zero. * <p> * Also supports {@link DeviceContext} to enrich a phone number during normalization if the optional NDC is missing. * </p> */ @RequiredArgsConstructor @Component public class PhoneNumberNormalizerImpl implements PhoneNumberNormalizer { private static final Logger LOGGER = LoggerFactory.getLogger(PhoneNumberNormalizerImpl.class); /** * Storage for {@link PhoneNumberNormalizer#setFallbackRegionCode(String)} */ private String fallbackRegionCode = null; @Override public void setFallbackRegionCode(String fallBackRegionCode) { if (PhoneLibWrapper.getCountryCodeForRegion(fallBackRegionCode) > 0) { this.fallbackRegionCode = fallBackRegionCode; } else { this.fallbackRegionCode = null; //invalid region code! } } /** * Fallback normalization within the number-plan of the fallback region. * @param number the original number to be normalized * @param dialableNumber the original number reduced to dialable digits * @return E164 formatted phone number or at least a dialable version of the number * * @see PhoneNumberNormalizer#setFallbackRegionCode(String) * @see PhoneNumberNormalizer#normalizePhoneNumber(String, String) */ private String fallbackNormalizationFromDeviceContextToDefaultRegionCode(String number, String dialableNumber) { if (this.fallbackRegionCode == null) { LOGGER.debug("Fallback Region was set!"); return dialableNumber; } else { return this.normalizePhoneNumber(number, this.fallbackRegionCode); } } /** * Uses wrapper of PhoneLib to identify if special rules apply for normalization.<br/> * Using device context for enriching the number make it normalizable to E164 format if NDC is optional in the used number plan, but not used in the phone number to be normalized. * @param wrapper instanced wrapper of PhoneLib * @param deviceContext information like CC, NDC and {@link de.telekom.phonenumbernormalizer.dto.DeviceContextLineType} from which the number is dialled * @return E164 formatted phone number or dialable version of it or null */ private String normalize(PhoneLibWrapper wrapper, DeviceContext deviceContext) { // international prefix has been added by PhoneLib even if it's not valid in the number plan. if (wrapper == null) { LOGGER.debug("PhoneLipWrapper was not initialized"); return null; } if (wrapper.getSemiNormalizedNumber() == null) { return wrapper.getDialableNumber(); } NumberPlan numberplan = null; if (deviceContext != null) { numberplan = NumberPlanFactory.INSTANCE.getNumberPlan(deviceContext.getLineType(), deviceContext.getCountryCode()); } if (wrapper.isShortNumber(numberplan)) { //if it is a short number, we can't add area code nor country code, so returning the dialable. return wrapper.getDialableNumber(); } if (wrapper.hasRegionNationalAccessCode() && deviceContext != null) { //Number plan is using a NationalPrefix aka Trunc Code ... so we could add Area Code if not included in the number. return wrapper.extendNumberByDefaultAreaCodeAndCountryCode(wrapper.getNationalAccessCode(), deviceContext.getNationalDestinationCode()); } // Number plan is not using NationalPrefix aka Trunc Code ... its also not a short number, so country code can be added:
return wrapper.getE164Formatted();
} @Override public String normalizePhoneNumber(String number, String regionCode) { PhoneLibWrapper wrapper = new PhoneLibWrapper(number, regionCode); if (wrapper.getSemiNormalizedNumber() == null) { return wrapper.getDialableNumber(); } if (wrapper.isShortNumber()) { //if it is a short number, we can't add area code nor country code, so returning the dialable. return wrapper.getDialableNumber(); } // international prefix is added by the lib even if it's not valid in the number plan. //checking if the input number is equal to the nationalNumber based on number plan and trunc code logic. boolean hasNoCCAndNoNAC = wrapper.hasNoCountryCodeNorNationalAccessCode(); LOGGER.debug("Number has no CC and no NAC: {}.", hasNoCCAndNoNAC); //if the number is definitely a short number or needs an area code but does not have it, we do not add the country code. return (hasNoCCAndNoNAC) ? wrapper.getDialableNumber() : wrapper.getE164Formatted(); } @Override public String normalizePhoneNumber(String number, DeviceContext deviceContext) { // checking if the number has a special format or is not valid at all. PhoneLibWrapper normalizerPhoneNumber = new PhoneLibWrapper(number, null); if (! normalizerPhoneNumber.isNormalizingTried()) { return normalizerPhoneNumber.getDialableNumber(); } DeviceContext normalizedDeviceContext = DeviceContextMapper.normalized(deviceContext); if (!normalizedDeviceContext.getCountryCode().equals(DeviceContext.UNKNOWN_VALUE)) { String regionCode = PhoneLibWrapper.getRegionCodeForCountryCode(normalizedDeviceContext.getCountryCode()); // now working again with the region code normalizerPhoneNumber = new PhoneLibWrapper(number, regionCode); if (!normalizedDeviceContext.getNationalDestinationCode().equals(DeviceContext.UNKNOWN_VALUE)) { // Number needs normalization: return normalize(normalizerPhoneNumber, deviceContext); } // Device Context with CountryCode but without AreaCode ... if (!(PhoneLibWrapper.UNKNOWN_REGIONCODE.equals(regionCode))) { return this.normalizePhoneNumber(number, regionCode); } } LOGGER.debug("Normalization based on DeviceContext did not work - falling back to normalization with fallback region."); return this.fallbackNormalizationFromDeviceContextToDefaultRegionCode(number, normalizerPhoneNumber.getDialableNumber()); } }
src/main/java/de/telekom/phonenumbernormalizer/PhoneNumberNormalizerImpl.java
telekom-phonenumber-normalizer-9a1b7ad
[ { "filename": "src/main/java/de/telekom/phonenumbernormalizer/numberplans/PhoneLibWrapper.java", "retrieved_chunk": " try {\n this.semiNormalizedNumber = phoneUtil.parse(extendedNumber, regionCode);\n // after area code has been added, we can add the country code by the lib:\n return getE164Formatted();\n } catch (NumberParseException e) {\n LOGGER.warn(\"could not parse extended number: {}\", extendedNumber);\n LOGGER.debug(\"{}\", e.getMessage());\n return dialableNumber;\n }\n } else {", "score": 0.8807629942893982 }, { "filename": "src/main/java/de/telekom/phonenumbernormalizer/numberplans/PhoneLibWrapper.java", "retrieved_chunk": " } else {\n if (!isSpecialFormat(dialableNumber)) {\n // Number needs normalization:\n // international prefix is added by the lib even if it's not valid in the number plan.\n this.isNormalizingTried = true;\n this.semiNormalizedNumber = PhoneLibWrapper.parseNumber(dialableNumber, regionCode);\n }\n }\n }\n }", "score": 0.8719232082366943 }, { "filename": "src/main/java/de/telekom/phonenumbernormalizer/numberplans/PhoneLibWrapper.java", "retrieved_chunk": " * @param defaultNationalDestinationCode the NDC to be added depending on the use telephone line origination.\n * @return if possible a E164 formatted representation or just the diallable representation of the given number.\n *\n * @see PhoneLibWrapper#PhoneLibWrapper(String, String)\n */\n public String extendNumberByDefaultAreaCodeAndCountryCode(String nationalAccessCode, String defaultNationalDestinationCode) {\n String nationalPhoneNumberWithoutNationalAccessCode = this.getNationalPhoneNumberWithoutNationalAccessCode();\n //if the dialableNumber is same as the national Number, Without NationalPrefix, then there is no NDC, so it needs to be added.\n if (Objects.equals(dialableNumber, nationalPhoneNumberWithoutNationalAccessCode)) {\n String extendedNumber = nationalAccessCode + defaultNationalDestinationCode + nationalPhoneNumberWithoutNationalAccessCode;", "score": 0.864497184753418 }, { "filename": "src/main/java/de/telekom/phonenumbernormalizer/numberplans/PhoneLibWrapper.java", "retrieved_chunk": " *\n * @see PhoneLibWrapper#PhoneLibWrapper(String, String)\n */\n public String getE164Formatted() {\n return phoneUtil.format(this.semiNormalizedNumber, PhoneNumberUtil.PhoneNumberFormat.E164);\n }\n /**\n * If we know the given region for the given number {@link PhoneLibWrapper#hasRegionNationalAccessCode()}, this method checks if the given number does not start with a NAC nor a CC,\n * so we could permanently add a default NDC and NAC to the given number and for this new value the method directly return a E164 formatted representation.\n * @param nationalAccessCode the NAC to be added e.g. for Germany it would be \"0\"", "score": 0.8563281297683716 }, { "filename": "src/main/java/de/telekom/phonenumbernormalizer/PhoneNumberAreaLabelImpl.java", "retrieved_chunk": " try {\n Phonenumber.PhoneNumber pn = phoneUtil.parse(e164number, \"\");\n Optional<String> locationName = Optional.empty();\n if (pn!=null) {\n String regionCode=phoneUtil.getRegionCodeForCountryCode(pn.getCountryCode());\n locationName = this.getLocationByNationalNumberAndRegionCode(String.valueOf(pn.getNationalNumber()), regionCode);\n if (locationName.isEmpty()) {\n return this.getCountryNameByCountryCode(String.valueOf(pn.getCountryCode()));\n }\n }", "score": 0.8547332882881165 } ]
java
return wrapper.getE164Formatted();
package com.rosymaple.hitindication.latesthits; import com.rosymaple.hitindication.config.HitIndicatorClientConfigs; import net.minecraft.client.Minecraft; import net.minecraft.world.entity.Entity; import net.minecraftforge.api.distmarker.Dist; import net.minecraftforge.api.distmarker.OnlyIn; import java.util.ArrayList; import java.util.UUID; @OnlyIn(Dist.CLIENT) public class ClientLatestHits { public static ArrayList<HitIndicator> latestHitIndicators = new ArrayList<>(); public static HitMarker currentHitMarker = null; public static void addHitIndicator(double x, double y, double z, int type, int damagePercent, boolean hasNegativeEffects) { HitIndicatorType hitIndicatorType = HitIndicatorType.fromInt(type); if(!HitIndicatorClientConfigs.EnableHitIndication.get()) return; if(hitIndicatorType == HitIndicatorType.BLUE && !HitIndicatorClientConfigs.ShowBlueIndicators.get()) return; if(hasNegativeEffects && !HitIndicatorClientConfigs.DisplayHitsFromNegativePotions.get()) return; latestHitIndicators.add(new HitIndicator(x, y, z, hitIndicatorType, damagePercent)); if(HitIndicatorClientConfigs.MaxIndicatorCount.get() > 0 && latestHitIndicators.size() > HitIndicatorClientConfigs.MaxIndicatorCount.get()) latestHitIndicators.remove(0); } public static void setHitMarker(int type) { HitMarkerType hitMarkerType = HitMarkerType.fromInt(type); if(!HitIndicatorClientConfigs.EnableHitMarkers.get()) return; currentHitMarker = new HitMarker(hitMarkerType); } public static void tick() { for(int i = latestHitIndicators.size()-1; i >= 0; i--) { HitIndicator hitIndicator = latestHitIndicators.get(i); hitIndicator.tick(); if
(hitIndicator.expired()) latestHitIndicators.remove(i);
} if(currentHitMarker != null) { currentHitMarker.tick(); if(currentHitMarker.expired()) currentHitMarker = null; } } public static void clear() { latestHitIndicators.clear(); currentHitMarker = null; } }
src/main/java/com/rosymaple/hitindication/latesthits/ClientLatestHits.java
TheHamester-HitIndicator-261bcb9
[ { "filename": "src/main/java/com/rosymaple/hitindication/latesthits/HitMarker.java", "retrieved_chunk": "package com.rosymaple.hitindication.latesthits;\npublic class HitMarker {\n private HitMarkerType hitMarkerType;\n private int lifetime;\n public HitMarker(HitMarkerType hitIndicatorType) {\n this.hitMarkerType = hitIndicatorType;\n lifetime = 9;\n }\n public void tick() {\n lifetime--;", "score": 0.9057537317276001 }, { "filename": "src/main/java/com/rosymaple/hitindication/latesthits/HitIndicator.java", "retrieved_chunk": " this.hitIndicatorType = hitIndicatorType;\n this.damagePercent = damagePercent;\n lifetime = HitIndicatorClientConfigs.FadeRate.get();\n }\n public void tick() {\n lifetime--;\n }\n public boolean expired() {\n return lifetime <= 0;\n }", "score": 0.8677418231964111 }, { "filename": "src/main/java/com/rosymaple/hitindication/event/RenderEvents.java", "retrieved_chunk": " for(HitIndicator hit : ClientLatestHits.latestHitIndicators) {\n drawIndicator(event.getMatrixStack(), hit, screenMiddleX, screenMiddleY, playerPos, lookVec);\n }\n if(ClientLatestHits.currentHitMarker != null)\n drawHitMarker(event.getMatrixStack(), ClientLatestHits.currentHitMarker, screenMiddleX, screenMiddleY);\n }\n private static void drawHitMarker(PoseStack stack, HitMarker hitMarker, int screenMiddleX, int screenMiddleY) {\n float opacity = hitMarker.getType() == HitMarkerType.CRIT ? 30 : 60;\n opacity /= 100.0f;\n bindMarkerTexture(hitMarker.getType(), hitMarker.getLifeTime());", "score": 0.8416744470596313 }, { "filename": "src/main/java/com/rosymaple/hitindication/latesthits/HitMarker.java", "retrieved_chunk": " }\n public boolean expired() {\n return lifetime <= 0;\n }\n public int getLifeTime() { return lifetime; }\n public HitMarkerType getType() { return hitMarkerType; }\n}", "score": 0.8375996947288513 }, { "filename": "src/main/java/com/rosymaple/hitindication/event/ClientEvents.java", "retrieved_chunk": " public static void onTick(TickEvent.ClientTickEvent event) {\n if(event.phase == TickEvent.Phase.END)\n return;\n ClientLatestHits.tick();\n }\n}", "score": 0.8329112529754639 } ]
java
(hitIndicator.expired()) latestHitIndicators.remove(i);
/* * Copyright © 2023 Deutsche Telekom AG (opensource@telekom.de) * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ package de.telekom.phonenumbernormalizer; import de.telekom.phonenumbernormalizer.dto.DeviceContext; import de.telekom.phonenumbernormalizer.dto.DeviceContextMapper; import de.telekom.phonenumbernormalizer.numberplans.PhoneLibWrapper; import de.telekom.phonenumbernormalizer.numberplans.NumberPlan; import de.telekom.phonenumbernormalizer.numberplans.NumberPlanFactory; import lombok.RequiredArgsConstructor; import org.slf4j.Logger; import org.slf4j.LoggerFactory; import org.springframework.stereotype.Component; /** * Concrete implementation of {@link PhoneNumberNormalizer} using {@link PhoneLibWrapper} to normalize a number by mitigating some inaccuracies when it comes to number plans of optional NDC and NAC as zero. * <p> * Also supports {@link DeviceContext} to enrich a phone number during normalization if the optional NDC is missing. * </p> */ @RequiredArgsConstructor @Component public class PhoneNumberNormalizerImpl implements PhoneNumberNormalizer { private static final Logger LOGGER = LoggerFactory.getLogger(PhoneNumberNormalizerImpl.class); /** * Storage for {@link PhoneNumberNormalizer#setFallbackRegionCode(String)} */ private String fallbackRegionCode = null; @Override public void setFallbackRegionCode(String fallBackRegionCode) { if (PhoneLibWrapper.getCountryCodeForRegion(fallBackRegionCode) > 0) { this.fallbackRegionCode = fallBackRegionCode; } else { this.fallbackRegionCode = null; //invalid region code! } } /** * Fallback normalization within the number-plan of the fallback region. * @param number the original number to be normalized * @param dialableNumber the original number reduced to dialable digits * @return E164 formatted phone number or at least a dialable version of the number * * @see PhoneNumberNormalizer#setFallbackRegionCode(String) * @see PhoneNumberNormalizer#normalizePhoneNumber(String, String) */ private String fallbackNormalizationFromDeviceContextToDefaultRegionCode(String number, String dialableNumber) { if (this.fallbackRegionCode == null) { LOGGER.debug("Fallback Region was set!"); return dialableNumber; } else { return this.normalizePhoneNumber(number, this.fallbackRegionCode); } } /** * Uses wrapper of PhoneLib to identify if special rules apply for normalization.<br/> * Using device context for enriching the number make it normalizable to E164 format if NDC is optional in the used number plan, but not used in the phone number to be normalized. * @param wrapper instanced wrapper of PhoneLib * @param deviceContext information like CC, NDC and {@link de.telekom.phonenumbernormalizer.dto.DeviceContextLineType} from which the number is dialled * @return E164 formatted phone number or dialable version of it or null */ private String normalize(PhoneLibWrapper wrapper, DeviceContext deviceContext) { // international prefix has been added by PhoneLib even if it's not valid in the number plan. if (wrapper == null) { LOGGER.debug("PhoneLipWrapper was not initialized"); return null; } if (wrapper.getSemiNormalizedNumber() == null) { return wrapper.getDialableNumber(); } NumberPlan numberplan = null; if (deviceContext != null) { numberplan = NumberPlanFactory.INSTANCE.getNumberPlan(deviceContext.getLineType(), deviceContext.getCountryCode()); } if (wrapper.isShortNumber(numberplan)) { //if it is a short number, we can't add area code nor country code, so returning the dialable. return wrapper.getDialableNumber(); } if (wrapper.hasRegionNationalAccessCode() && deviceContext != null) { //Number plan is using a NationalPrefix aka Trunc Code ... so we could add Area Code if not included in the number. return wrapper.extendNumberByDefaultAreaCodeAndCountryCode(wrapper.getNationalAccessCode(), deviceContext.getNationalDestinationCode()); } // Number plan is not using NationalPrefix aka Trunc Code ... its also not a short number, so country code can be added: return wrapper.getE164Formatted(); } @Override public String normalizePhoneNumber(String number, String regionCode) { PhoneLibWrapper wrapper = new PhoneLibWrapper(number, regionCode); if (wrapper.getSemiNormalizedNumber() == null) { return wrapper.getDialableNumber(); } if (wrapper.isShortNumber()) { //if it is a short number, we can't add area code nor country code, so returning the dialable. return wrapper.getDialableNumber(); } // international prefix is added by the lib even if it's not valid in the number plan. //checking if the input number is equal to the nationalNumber based on number plan and trunc code logic.
boolean hasNoCCAndNoNAC = wrapper.hasNoCountryCodeNorNationalAccessCode();
LOGGER.debug("Number has no CC and no NAC: {}.", hasNoCCAndNoNAC); //if the number is definitely a short number or needs an area code but does not have it, we do not add the country code. return (hasNoCCAndNoNAC) ? wrapper.getDialableNumber() : wrapper.getE164Formatted(); } @Override public String normalizePhoneNumber(String number, DeviceContext deviceContext) { // checking if the number has a special format or is not valid at all. PhoneLibWrapper normalizerPhoneNumber = new PhoneLibWrapper(number, null); if (! normalizerPhoneNumber.isNormalizingTried()) { return normalizerPhoneNumber.getDialableNumber(); } DeviceContext normalizedDeviceContext = DeviceContextMapper.normalized(deviceContext); if (!normalizedDeviceContext.getCountryCode().equals(DeviceContext.UNKNOWN_VALUE)) { String regionCode = PhoneLibWrapper.getRegionCodeForCountryCode(normalizedDeviceContext.getCountryCode()); // now working again with the region code normalizerPhoneNumber = new PhoneLibWrapper(number, regionCode); if (!normalizedDeviceContext.getNationalDestinationCode().equals(DeviceContext.UNKNOWN_VALUE)) { // Number needs normalization: return normalize(normalizerPhoneNumber, deviceContext); } // Device Context with CountryCode but without AreaCode ... if (!(PhoneLibWrapper.UNKNOWN_REGIONCODE.equals(regionCode))) { return this.normalizePhoneNumber(number, regionCode); } } LOGGER.debug("Normalization based on DeviceContext did not work - falling back to normalization with fallback region."); return this.fallbackNormalizationFromDeviceContextToDefaultRegionCode(number, normalizerPhoneNumber.getDialableNumber()); } }
src/main/java/de/telekom/phonenumbernormalizer/PhoneNumberNormalizerImpl.java
telekom-phonenumber-normalizer-9a1b7ad
[ { "filename": "src/main/java/de/telekom/phonenumbernormalizer/numberplans/PhoneLibWrapper.java", "retrieved_chunk": " } else {\n if (!isSpecialFormat(dialableNumber)) {\n // Number needs normalization:\n // international prefix is added by the lib even if it's not valid in the number plan.\n this.isNormalizingTried = true;\n this.semiNormalizedNumber = PhoneLibWrapper.parseNumber(dialableNumber, regionCode);\n }\n }\n }\n }", "score": 0.8813831806182861 }, { "filename": "src/main/java/de/telekom/phonenumbernormalizer/numberplans/PhoneLibWrapper.java", "retrieved_chunk": " * If we have a plain national number based on regions number plan and potential NAC logic.\n * <p>\n * For a number plan without NAC logic, it will always return false!\n * </p>\n * @return if given number could have CC and NAC, but does not have any of them.\n */\n public boolean hasNoCountryCodeNorNationalAccessCode() {\n // if given number has no NAC and no CC, it equals national phone number (without NAC).\n if (! Objects.equals(dialableNumber, this.getNationalPhoneNumberWithoutNationalAccessCode())) {\n return false;", "score": 0.877981424331665 }, { "filename": "src/main/java/de/telekom/phonenumbernormalizer/numberplans/PhoneLibWrapper.java", "retrieved_chunk": " *\n * @return national number without NAC, but any other leading zero.\n *\n * @see PhoneLibWrapper#PhoneLibWrapper(String, String)\n * @see PhoneLibWrapper#getSemiNormalizedNumber()\n * @see PhoneLibWrapper#nationalPhoneNumberWithoutNationalPrefix(Phonenumber.PhoneNumber)\n */\n private String getNationalPhoneNumberWithoutNationalAccessCode() {\n return PhoneLibWrapper.nationalPhoneNumberWithoutNationalPrefix(this.semiNormalizedNumber);\n }", "score": 0.8671857118606567 }, { "filename": "src/main/java/de/telekom/phonenumbernormalizer/numberplans/PhoneLibWrapper.java", "retrieved_chunk": " try {\n this.semiNormalizedNumber = phoneUtil.parse(extendedNumber, regionCode);\n // after area code has been added, we can add the country code by the lib:\n return getE164Formatted();\n } catch (NumberParseException e) {\n LOGGER.warn(\"could not parse extended number: {}\", extendedNumber);\n LOGGER.debug(\"{}\", e.getMessage());\n return dialableNumber;\n }\n } else {", "score": 0.8565121293067932 }, { "filename": "src/main/java/de/telekom/phonenumbernormalizer/numberplans/PhoneLibWrapper.java", "retrieved_chunk": " }\n return metadata.getNationalPrefix();\n }\n /**\n * From PhoneLib, if a National Access Code is used before the National Destination Code in the given region\n * @return if given {@link PhoneLibWrapper#regionCode} is using NAC\n */\n public boolean hasRegionNationalAccessCode() {\n return metadata != null && metadata.hasNationalPrefix();\n }", "score": 0.8534061312675476 } ]
java
boolean hasNoCCAndNoNAC = wrapper.hasNoCountryCodeNorNationalAccessCode();
/* * Copyright © 2023 Deutsche Telekom AG (opensource@telekom.de) * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ package de.telekom.phonenumbernormalizer; import de.telekom.phonenumbernormalizer.dto.DeviceContext; import de.telekom.phonenumbernormalizer.dto.DeviceContextMapper; import de.telekom.phonenumbernormalizer.numberplans.PhoneLibWrapper; import de.telekom.phonenumbernormalizer.numberplans.NumberPlan; import de.telekom.phonenumbernormalizer.numberplans.NumberPlanFactory; import lombok.RequiredArgsConstructor; import org.slf4j.Logger; import org.slf4j.LoggerFactory; import org.springframework.stereotype.Component; /** * Concrete implementation of {@link PhoneNumberNormalizer} using {@link PhoneLibWrapper} to normalize a number by mitigating some inaccuracies when it comes to number plans of optional NDC and NAC as zero. * <p> * Also supports {@link DeviceContext} to enrich a phone number during normalization if the optional NDC is missing. * </p> */ @RequiredArgsConstructor @Component public class PhoneNumberNormalizerImpl implements PhoneNumberNormalizer { private static final Logger LOGGER = LoggerFactory.getLogger(PhoneNumberNormalizerImpl.class); /** * Storage for {@link PhoneNumberNormalizer#setFallbackRegionCode(String)} */ private String fallbackRegionCode = null; @Override public void setFallbackRegionCode(String fallBackRegionCode) { if (PhoneLibWrapper.getCountryCodeForRegion(fallBackRegionCode) > 0) { this.fallbackRegionCode = fallBackRegionCode; } else { this.fallbackRegionCode = null; //invalid region code! } } /** * Fallback normalization within the number-plan of the fallback region. * @param number the original number to be normalized * @param dialableNumber the original number reduced to dialable digits * @return E164 formatted phone number or at least a dialable version of the number * * @see PhoneNumberNormalizer#setFallbackRegionCode(String) * @see PhoneNumberNormalizer#normalizePhoneNumber(String, String) */ private String fallbackNormalizationFromDeviceContextToDefaultRegionCode(String number, String dialableNumber) { if (this.fallbackRegionCode == null) { LOGGER.debug("Fallback Region was set!"); return dialableNumber; } else { return this.normalizePhoneNumber(number, this.fallbackRegionCode); } } /** * Uses wrapper of PhoneLib to identify if special rules apply for normalization.<br/> * Using device context for enriching the number make it normalizable to E164 format if NDC is optional in the used number plan, but not used in the phone number to be normalized. * @param wrapper instanced wrapper of PhoneLib * @param deviceContext information like CC, NDC and {@link de.telekom.phonenumbernormalizer.dto.DeviceContextLineType} from which the number is dialled * @return E164 formatted phone number or dialable version of it or null */ private String normalize(PhoneLibWrapper wrapper, DeviceContext deviceContext) { // international prefix has been added by PhoneLib even if it's not valid in the number plan. if (wrapper == null) { LOGGER.debug("PhoneLipWrapper was not initialized"); return null; } if (wrapper.getSemiNormalizedNumber() == null) { return wrapper.getDialableNumber(); } NumberPlan numberplan = null; if (deviceContext != null) { numberplan = NumberPlanFactory.INSTANCE.getNumberPlan(deviceContext.getLineType(), deviceContext.getCountryCode()); } if (wrapper.isShortNumber(numberplan)) { //if it is a short number, we can't add area code nor country code, so returning the dialable. return wrapper.getDialableNumber(); } if (wrapper.hasRegionNationalAccessCode() && deviceContext != null) { //Number plan is using a NationalPrefix aka Trunc Code ... so we could add Area Code if not included in the number. return wrapper.extendNumberByDefaultAreaCodeAndCountryCode(wrapper.getNationalAccessCode(), deviceContext.getNationalDestinationCode()); } // Number plan is not using NationalPrefix aka Trunc Code ... its also not a short number, so country code can be added: return wrapper.getE164Formatted(); } @Override public String normalizePhoneNumber(String number, String regionCode) { PhoneLibWrapper wrapper = new PhoneLibWrapper(number, regionCode); if (wrapper.getSemiNormalizedNumber() == null) { return wrapper.getDialableNumber(); } if (wrapper.isShortNumber()) { //if it is a short number, we can't add area code nor country code, so returning the dialable. return wrapper.getDialableNumber(); } // international prefix is added by the lib even if it's not valid in the number plan. //checking if the input number is equal to the nationalNumber based on number plan and trunc code logic. boolean hasNoCCAndNoNAC = wrapper.hasNoCountryCodeNorNationalAccessCode(); LOGGER.debug("Number has no CC and no NAC: {}.", hasNoCCAndNoNAC); //if the number is definitely a short number or needs an area code but does not have it, we do not add the country code. return (hasNoCCAndNoNAC) ? wrapper.getDialableNumber() : wrapper.getE164Formatted(); } @Override public String normalizePhoneNumber(String number, DeviceContext deviceContext) { // checking if the number has a special format or is not valid at all. PhoneLibWrapper normalizerPhoneNumber = new PhoneLibWrapper(number, null); if (! normalizerPhoneNumber.isNormalizingTried()) { return normalizerPhoneNumber.getDialableNumber(); } DeviceContext normalizedDeviceContext = DeviceContextMapper.normalized(deviceContext); if (!normalizedDeviceContext.getCountryCode().equals(DeviceContext.UNKNOWN_VALUE)) { String regionCode = PhoneLibWrapper.getRegionCodeForCountryCode(normalizedDeviceContext.getCountryCode()); // now working again with the region code normalizerPhoneNumber = new PhoneLibWrapper(number, regionCode); if (
!normalizedDeviceContext.getNationalDestinationCode().equals(DeviceContext.UNKNOWN_VALUE)) {
// Number needs normalization: return normalize(normalizerPhoneNumber, deviceContext); } // Device Context with CountryCode but without AreaCode ... if (!(PhoneLibWrapper.UNKNOWN_REGIONCODE.equals(regionCode))) { return this.normalizePhoneNumber(number, regionCode); } } LOGGER.debug("Normalization based on DeviceContext did not work - falling back to normalization with fallback region."); return this.fallbackNormalizationFromDeviceContextToDefaultRegionCode(number, normalizerPhoneNumber.getDialableNumber()); } }
src/main/java/de/telekom/phonenumbernormalizer/PhoneNumberNormalizerImpl.java
telekom-phonenumber-normalizer-9a1b7ad
[ { "filename": "src/main/java/de/telekom/phonenumbernormalizer/numberplans/PhoneLibWrapper.java", "retrieved_chunk": " } else {\n if (!isSpecialFormat(dialableNumber)) {\n // Number needs normalization:\n // international prefix is added by the lib even if it's not valid in the number plan.\n this.isNormalizingTried = true;\n this.semiNormalizedNumber = PhoneLibWrapper.parseNumber(dialableNumber, regionCode);\n }\n }\n }\n }", "score": 0.9134132862091064 }, { "filename": "src/main/java/de/telekom/phonenumbernormalizer/numberplans/PhoneLibWrapper.java", "retrieved_chunk": " try {\n this.semiNormalizedNumber = phoneUtil.parse(extendedNumber, regionCode);\n // after area code has been added, we can add the country code by the lib:\n return getE164Formatted();\n } catch (NumberParseException e) {\n LOGGER.warn(\"could not parse extended number: {}\", extendedNumber);\n LOGGER.debug(\"{}\", e.getMessage());\n return dialableNumber;\n }\n } else {", "score": 0.8952082395553589 }, { "filename": "src/main/java/de/telekom/phonenumbernormalizer/PhoneNumberNormalizer.java", "retrieved_chunk": " */\n String normalizePhoneNumber(String number, DeviceContext deviceContext);\n /**\n * Normalizes the number using PhoneLib with some additions to compensate.\n * <p>\n * Not as powerful as {@link PhoneNumberNormalizer#normalizePhoneNumber(String, DeviceContext)}, because no default NDC can be set.\n * </p>\n * @param number plain number to normalize\n * @param regionCode ISO2 code of the country, which number-plan is used for normalization\n * @return E164 formatted phone number or at least a dialable version of the number", "score": 0.8910413980484009 }, { "filename": "src/main/java/de/telekom/phonenumbernormalizer/numberplans/PhoneLibWrapper.java", "retrieved_chunk": " String dialableNumber;\n /**\n * The given number normalized with PhoneLib, risking we get a incorrect normalization\n *\n * @see PhoneLibWrapper#PhoneLibWrapper(String, String)\n * @see PhoneLibWrapper#isNormalizingTried()\n * @see PhoneLibWrapper#getSemiNormalizedNumber()\n */\n Phonenumber.PhoneNumber semiNormalizedNumber;\n /**", "score": 0.8864673972129822 }, { "filename": "src/main/java/de/telekom/phonenumbernormalizer/PhoneNumberNormalizer.java", "retrieved_chunk": " /**\n * Normalizes the number using PhoneLib with some additions to compensate.\n * <p>\n * Preferable to {@link PhoneNumberNormalizer#normalizePhoneNumber(String, String)}, because default NDC can be provided, so that more compensation for generating a valid E164 can be done.\n * </p>\n * @param number plain number to normalize\n * @param deviceContext information like CC, NDC and {@link de.telekom.phonenumbernormalizer.dto.DeviceContextLineType} from which the number is dialled\n * @return E164 formatted phone number or at least a dialable version of the number\n *\n * @see PhoneNumberNormalizer#setFallbackRegionCode(String)", "score": 0.87656170129776 } ]
java
!normalizedDeviceContext.getNationalDestinationCode().equals(DeviceContext.UNKNOWN_VALUE)) {
package net.chauvedev.woodencog.mixin; import com.simibubi.create.content.kinetics.KineticNetwork; import com.simibubi.create.content.kinetics.base.KineticBlockEntity; import com.simibubi.create.content.kinetics.crusher.CrushingWheelBlockEntity; import com.simibubi.create.content.kinetics.transmission.ClutchBlockEntity; import net.chauvedev.woodencog.WoodenCog; import net.chauvedev.woodencog.capability.MachineCapacity; import net.chauvedev.woodencog.capability.MachineCapacityProvider; import net.chauvedev.woodencog.core.MachineCapacityEntry; import net.chauvedev.woodencog.core.MachineCapacityStorage; import net.minecraft.core.BlockPos; import net.minecraft.core.particles.ParticleTypes; import net.minecraft.nbt.CompoundTag; import net.minecraft.sounds.SoundEvent; import net.minecraft.sounds.SoundEvents; import net.minecraft.sounds.SoundSource; import net.minecraft.world.level.Level; import net.minecraft.world.level.block.Block; import net.minecraft.world.level.block.Blocks; import net.minecraft.world.level.block.entity.BlockEntity; import net.minecraft.world.level.block.entity.BlockEntityType; import net.minecraft.world.level.block.state.BlockState; import org.spongepowered.asm.mixin.Mixin; import org.spongepowered.asm.mixin.Shadow; import org.spongepowered.asm.mixin.injection.At; import org.spongepowered.asm.mixin.injection.Inject; import org.spongepowered.asm.mixin.injection.callback.CallbackInfo; import java.awt.*; import java.util.Objects; @Mixin(value = KineticBlockEntity.class, remap = false) public abstract class MixinKineticBlockEntity{ private boolean lifeBlockBlocked = false; private float lifeBlock = 20*10; @Shadow public abstract KineticNetwork getOrCreateNetwork(); @Shadow protected float speed; @Shadow protected float stress; @Shadow protected boolean overStressed; @Shadow public abstract void setSpeed(float speed); @Shadow protected float capacity; @Shadow public abstract float calculateAddedStressCapacity(); @Shadow public abstract float calculateStressApplied(); @Shadow protected float lastStressApplied; @Shadow protected float lastCapacityProvided; @Shadow public abstract void detachKinetics(); @Shadow public abstract float getSpeed(); boolean destroyed = false; @Inject( method = {"tick"}, at = {@At("HEAD")}, remap = false) public void tick(CallbackInfo ci){ KineticBlockEntity block = (KineticBlockEntity)((Object)this) ; try{ this.tickDamagedTick(block); }catch (Error error){ WoodenCog.LOGGER.debug("Error pendant le system tick"); error.printStackTrace(); } } public void tickDamagedTick(KineticBlockEntity block){ if (!MachineCapacityStorage.getInstance().active){ return; } if (MachineCapacityStorage.getInstance().isBlackListBlock(block)){ return; } MachineCapacityEntry config = MachineCapacityStorage.getInstance().getCapacity(block.getBlockState().getBlock()); MachineCapacity capacity = block.getCapability(MachineCapacityProvider.MACHINE_CAPACITY).resolve().get(); int chance = block.getLevel().random.nextInt(0,101); float left = config.durabilityMax - capacity.getDurability(); if (chance > (100-config.damageChance) && left > 0 && block.getSpeed() > 0){ int damage_for_speed = (int) getSpeed()/10; capacity.setDurability(capacity.getDurability()+damage_for_speed); } left = config.durabilityMax - capacity.getDurability(); if(left >= 10 && left < config.durabilityMax/2) { block.getLevel().addParticle(ParticleTypes.SMOKE,false,block.getBlockPos().getX()+0.5f,block.getBlockPos().getY()+0.5f,block.getBlockPos().getZ()+0.5f,0,0.01,0); } if (left<10){ if (getOrCreateNetwork()!=null){
if (!destroyed || !capacity.isDestroyed() || this.stress != Integer.MAX_VALUE){
destroyed = true; getOrCreateNetwork().updateCapacityFor(block,0); getOrCreateNetwork().updateStressFor(block, Integer.MAX_VALUE); getOrCreateNetwork().updateNetwork(); getOrCreateNetwork().sync(); capacity.setDestroyed(true); } } block.getLevel().addParticle(ParticleTypes.CAMPFIRE_SIGNAL_SMOKE,false,block.getBlockPos().getX()+0.5f,block.getBlockPos().getY()+0.5f,block.getBlockPos().getZ()+0.5f,0,0.01,0); } } }
src/main/java/net/chauvedev/woodencog/mixin/MixinKineticBlockEntity.java
Sparks-and-Rotation-woodencog-bb3d3ef
[ { "filename": "src/main/java/net/chauvedev/woodencog/mixin/MixinKineticBlock.java", "retrieved_chunk": " } else {\n capacity.setDurability(0);\n int durability_to_remove = (int)(capacity.getDurability() / 10) + 1;\n itemInHand.hurtAndBreak(durability_to_remove, pPlayer, (player -> player.broadcastBreakEvent(player.getUsedItemHand())));\n pLevel.playLocalSound(\n pPos.getX(),\n pPos.getY(),\n pPos.getZ(),\n SoundEvents.ANVIL_USE,\n SoundSource.BLOCKS,", "score": 0.8755189180374146 }, { "filename": "src/main/java/net/chauvedev/woodencog/capability/MachineCapacity.java", "retrieved_chunk": "package net.chauvedev.woodencog.capability;\nimport net.minecraft.nbt.CompoundTag;\nimport org.checkerframework.checker.units.qual.C;\npublic class MachineCapacity {\n private float machine_capacity = 0;\n private boolean destroyed = false;\n public float getDurability(){\n return this.machine_capacity;\n }\n public void setDurability(float var1){", "score": 0.8415712714195251 }, { "filename": "src/main/java/net/chauvedev/woodencog/mixin/MixinFanProcessing.java", "retrieved_chunk": " if(cap.getTemperature() <= 0F) {\n cap.setTemperature(0F);\n }\n } else if (type.equals(FanProcessing.Type.SPLASHING)) {\n cap.setTemperature(cap.getTemperature() - 5F);\n if(cap.getTemperature() <= 0F) {\n cap.setTemperature(0F);\n }\n }\n HeatingRecipe recipe = HeatingRecipe.getRecipe(inputStack);", "score": 0.8136787414550781 }, { "filename": "src/main/java/net/chauvedev/woodencog/config/WoodenCogCommonConfigs.java", "retrieved_chunk": " .define(\"durability\", 144000);\n DEFAULT_DAMAGE_CHANCE = BUILDER\n .comment(\"chance of machine getting damage (from 0 to 100 number over 100 are the same as 100)\")\n .define(\"chance\", 10);\n WEAR_BLACKLIST = BUILDER\n .comment(\"This list contains block that should not damage over time\")\n .defineList(\"blacklist\", List.of(\n \"create:shaft\",\n \"create:cogwheel\",\n \"create:large_cogwheel\",", "score": 0.8132657408714294 }, { "filename": "src/main/java/net/chauvedev/woodencog/WoodenCogEvents.java", "retrieved_chunk": " Block block = ((BlockItem) event.getObject().getItem()).getBlock();\n if (block instanceof KineticBlock){\n event.addCapability(MachineCapacityProvider.MACHINE_CAPACITY_KEY, new MachineCapacityProvider());\n }\n }\n }catch (Error error){\n WoodenCog.LOGGER.debug(\"-------ERROR---\");\n WoodenCog.LOGGER.error(\"Error found\",error);\n }\n }", "score": 0.8078428506851196 } ]
java
if (!destroyed || !capacity.isDestroyed() || this.stress != Integer.MAX_VALUE){
package net.chauvedev.woodencog.mixin; import com.simibubi.create.content.kinetics.KineticNetwork; import com.simibubi.create.content.kinetics.base.KineticBlockEntity; import com.simibubi.create.content.kinetics.crusher.CrushingWheelBlockEntity; import com.simibubi.create.content.kinetics.transmission.ClutchBlockEntity; import net.chauvedev.woodencog.WoodenCog; import net.chauvedev.woodencog.capability.MachineCapacity; import net.chauvedev.woodencog.capability.MachineCapacityProvider; import net.chauvedev.woodencog.core.MachineCapacityEntry; import net.chauvedev.woodencog.core.MachineCapacityStorage; import net.minecraft.core.BlockPos; import net.minecraft.core.particles.ParticleTypes; import net.minecraft.nbt.CompoundTag; import net.minecraft.sounds.SoundEvent; import net.minecraft.sounds.SoundEvents; import net.minecraft.sounds.SoundSource; import net.minecraft.world.level.Level; import net.minecraft.world.level.block.Block; import net.minecraft.world.level.block.Blocks; import net.minecraft.world.level.block.entity.BlockEntity; import net.minecraft.world.level.block.entity.BlockEntityType; import net.minecraft.world.level.block.state.BlockState; import org.spongepowered.asm.mixin.Mixin; import org.spongepowered.asm.mixin.Shadow; import org.spongepowered.asm.mixin.injection.At; import org.spongepowered.asm.mixin.injection.Inject; import org.spongepowered.asm.mixin.injection.callback.CallbackInfo; import java.awt.*; import java.util.Objects; @Mixin(value = KineticBlockEntity.class, remap = false) public abstract class MixinKineticBlockEntity{ private boolean lifeBlockBlocked = false; private float lifeBlock = 20*10; @Shadow public abstract KineticNetwork getOrCreateNetwork(); @Shadow protected float speed; @Shadow protected float stress; @Shadow protected boolean overStressed; @Shadow public abstract void setSpeed(float speed); @Shadow protected float capacity; @Shadow public abstract float calculateAddedStressCapacity(); @Shadow public abstract float calculateStressApplied(); @Shadow protected float lastStressApplied; @Shadow protected float lastCapacityProvided; @Shadow public abstract void detachKinetics(); @Shadow public abstract float getSpeed(); boolean destroyed = false; @Inject( method = {"tick"}, at = {@At("HEAD")}, remap = false) public void tick(CallbackInfo ci){ KineticBlockEntity block = (KineticBlockEntity)((Object)this) ; try{ this.tickDamagedTick(block); }catch (Error error){ WoodenCog.LOGGER.debug("Error pendant le system tick"); error.printStackTrace(); } } public void tickDamagedTick(KineticBlockEntity block){ if (!MachineCapacityStorage.getInstance().active){ return; } if (MachineCapacityStorage.getInstance().isBlackListBlock(block)){ return; } MachineCapacityEntry config = MachineCapacityStorage.getInstance().getCapacity(block.getBlockState().getBlock()); MachineCapacity capacity = block.getCapability(MachineCapacityProvider.MACHINE_CAPACITY).resolve().get(); int chance = block.getLevel().random.nextInt(0,101); float left = config.durabilityMax - capacity.getDurability(); if (chance > (100-config.damageChance) && left > 0 && block.getSpeed() > 0){ int damage_for_speed = (int) getSpeed()/10; capacity.setDurability(capacity.getDurability()+damage_for_speed); } left = config.durabilityMax - capacity.getDurability(); if(left >= 10 && left < config.durabilityMax/2) { block.getLevel().addParticle(ParticleTypes.SMOKE,false,block.getBlockPos().getX()+0.5f,block.getBlockPos().getY()+0.5f,block.getBlockPos().getZ()+0.5f,0,0.01,0); } if (left<10){ if (getOrCreateNetwork()!=null){ if (!destroyed || !capacity.isDestroyed() || this.stress != Integer.MAX_VALUE){ destroyed = true; getOrCreateNetwork().updateCapacityFor(block,0); getOrCreateNetwork().updateStressFor(block, Integer.MAX_VALUE); getOrCreateNetwork().updateNetwork(); getOrCreateNetwork().sync();
capacity.setDestroyed(true);
} } block.getLevel().addParticle(ParticleTypes.CAMPFIRE_SIGNAL_SMOKE,false,block.getBlockPos().getX()+0.5f,block.getBlockPos().getY()+0.5f,block.getBlockPos().getZ()+0.5f,0,0.01,0); } } }
src/main/java/net/chauvedev/woodencog/mixin/MixinKineticBlockEntity.java
Sparks-and-Rotation-woodencog-bb3d3ef
[ { "filename": "src/main/java/net/chauvedev/woodencog/core/MachineCapacityStorage.java", "retrieved_chunk": " public MachineCapacityEntry getCapacity(Block block){\n String full = block.getRegistryName().getNamespace()+\":\"+block.getRegistryName().getPath();\n try {\n CustomBlockConfig.BlockInformation info = CustomBlockConfig.registeredBlocks.get(full);\n return MachineCapacityEntry.createEntryBlock(false,\n info.durability,\n info.chance\n );\n } catch (NullPointerException e) {\n return MachineCapacityEntry.createEntryBlock(false,", "score": 0.8111191391944885 }, { "filename": "src/main/java/net/chauvedev/woodencog/core/MachineCapacityStorage.java", "retrieved_chunk": " public boolean active = true;\n public MachineCapacityStorage(){\n MachineCapacityStorage.INSTANCE = this;\n this.machineCapacityEntryMap_cache = new HashMap<>();\n }\n public boolean isBlackListBlock(BlockEntity blockEntity){\n Block block = blockEntity.getBlockState().getBlock();\n String full = block.getRegistryName().getNamespace()+\":\"+block.getRegistryName().getPath();\n return WoodenCogCommonConfigs.WEAR_BLACKLIST.get().contains(full);\n }", "score": 0.7960162162780762 }, { "filename": "src/main/java/net/chauvedev/woodencog/WoodenCogEvents.java", "retrieved_chunk": " Block block = ((BlockItem) event.getObject().getItem()).getBlock();\n if (block instanceof KineticBlock){\n event.addCapability(MachineCapacityProvider.MACHINE_CAPACITY_KEY, new MachineCapacityProvider());\n }\n }\n }catch (Error error){\n WoodenCog.LOGGER.debug(\"-------ERROR---\");\n WoodenCog.LOGGER.error(\"Error found\",error);\n }\n }", "score": 0.7850953340530396 }, { "filename": "src/main/java/net/chauvedev/woodencog/mixin/MixinKineticBlock.java", "retrieved_chunk": " } else {\n capacity.setDurability(0);\n int durability_to_remove = (int)(capacity.getDurability() / 10) + 1;\n itemInHand.hurtAndBreak(durability_to_remove, pPlayer, (player -> player.broadcastBreakEvent(player.getUsedItemHand())));\n pLevel.playLocalSound(\n pPos.getX(),\n pPos.getY(),\n pPos.getZ(),\n SoundEvents.ANVIL_USE,\n SoundSource.BLOCKS,", "score": 0.7819342613220215 }, { "filename": "src/main/java/net/chauvedev/woodencog/capability/MachineCapacity.java", "retrieved_chunk": "package net.chauvedev.woodencog.capability;\nimport net.minecraft.nbt.CompoundTag;\nimport org.checkerframework.checker.units.qual.C;\npublic class MachineCapacity {\n private float machine_capacity = 0;\n private boolean destroyed = false;\n public float getDurability(){\n return this.machine_capacity;\n }\n public void setDurability(float var1){", "score": 0.7780029773712158 } ]
java
capacity.setDestroyed(true);
package net.chauvedev.woodencog.mixin; import com.simibubi.create.content.kinetics.KineticNetwork; import com.simibubi.create.content.kinetics.base.KineticBlockEntity; import com.simibubi.create.content.kinetics.crusher.CrushingWheelBlockEntity; import com.simibubi.create.content.kinetics.transmission.ClutchBlockEntity; import net.chauvedev.woodencog.WoodenCog; import net.chauvedev.woodencog.capability.MachineCapacity; import net.chauvedev.woodencog.capability.MachineCapacityProvider; import net.chauvedev.woodencog.core.MachineCapacityEntry; import net.chauvedev.woodencog.core.MachineCapacityStorage; import net.minecraft.core.BlockPos; import net.minecraft.core.particles.ParticleTypes; import net.minecraft.nbt.CompoundTag; import net.minecraft.sounds.SoundEvent; import net.minecraft.sounds.SoundEvents; import net.minecraft.sounds.SoundSource; import net.minecraft.world.level.Level; import net.minecraft.world.level.block.Block; import net.minecraft.world.level.block.Blocks; import net.minecraft.world.level.block.entity.BlockEntity; import net.minecraft.world.level.block.entity.BlockEntityType; import net.minecraft.world.level.block.state.BlockState; import org.spongepowered.asm.mixin.Mixin; import org.spongepowered.asm.mixin.Shadow; import org.spongepowered.asm.mixin.injection.At; import org.spongepowered.asm.mixin.injection.Inject; import org.spongepowered.asm.mixin.injection.callback.CallbackInfo; import java.awt.*; import java.util.Objects; @Mixin(value = KineticBlockEntity.class, remap = false) public abstract class MixinKineticBlockEntity{ private boolean lifeBlockBlocked = false; private float lifeBlock = 20*10; @Shadow public abstract KineticNetwork getOrCreateNetwork(); @Shadow protected float speed; @Shadow protected float stress; @Shadow protected boolean overStressed; @Shadow public abstract void setSpeed(float speed); @Shadow protected float capacity; @Shadow public abstract float calculateAddedStressCapacity(); @Shadow public abstract float calculateStressApplied(); @Shadow protected float lastStressApplied; @Shadow protected float lastCapacityProvided; @Shadow public abstract void detachKinetics(); @Shadow public abstract float getSpeed(); boolean destroyed = false; @Inject( method = {"tick"}, at = {@At("HEAD")}, remap = false) public void tick(CallbackInfo ci){ KineticBlockEntity block = (KineticBlockEntity)((Object)this) ; try{ this.tickDamagedTick(block); }catch (Error error){ WoodenCog.LOGGER.debug("Error pendant le system tick"); error.printStackTrace(); } } public void tickDamagedTick(KineticBlockEntity block){ if (!MachineCapacityStorage.getInstance().active){ return; } if (MachineCapacityStorage.getInstance().isBlackListBlock(block)){ return; }
MachineCapacityEntry config = MachineCapacityStorage.getInstance().getCapacity(block.getBlockState().getBlock());
MachineCapacity capacity = block.getCapability(MachineCapacityProvider.MACHINE_CAPACITY).resolve().get(); int chance = block.getLevel().random.nextInt(0,101); float left = config.durabilityMax - capacity.getDurability(); if (chance > (100-config.damageChance) && left > 0 && block.getSpeed() > 0){ int damage_for_speed = (int) getSpeed()/10; capacity.setDurability(capacity.getDurability()+damage_for_speed); } left = config.durabilityMax - capacity.getDurability(); if(left >= 10 && left < config.durabilityMax/2) { block.getLevel().addParticle(ParticleTypes.SMOKE,false,block.getBlockPos().getX()+0.5f,block.getBlockPos().getY()+0.5f,block.getBlockPos().getZ()+0.5f,0,0.01,0); } if (left<10){ if (getOrCreateNetwork()!=null){ if (!destroyed || !capacity.isDestroyed() || this.stress != Integer.MAX_VALUE){ destroyed = true; getOrCreateNetwork().updateCapacityFor(block,0); getOrCreateNetwork().updateStressFor(block, Integer.MAX_VALUE); getOrCreateNetwork().updateNetwork(); getOrCreateNetwork().sync(); capacity.setDestroyed(true); } } block.getLevel().addParticle(ParticleTypes.CAMPFIRE_SIGNAL_SMOKE,false,block.getBlockPos().getX()+0.5f,block.getBlockPos().getY()+0.5f,block.getBlockPos().getZ()+0.5f,0,0.01,0); } } }
src/main/java/net/chauvedev/woodencog/mixin/MixinKineticBlockEntity.java
Sparks-and-Rotation-woodencog-bb3d3ef
[ { "filename": "src/main/java/net/chauvedev/woodencog/core/MachineCapacityStorage.java", "retrieved_chunk": " public boolean active = true;\n public MachineCapacityStorage(){\n MachineCapacityStorage.INSTANCE = this;\n this.machineCapacityEntryMap_cache = new HashMap<>();\n }\n public boolean isBlackListBlock(BlockEntity blockEntity){\n Block block = blockEntity.getBlockState().getBlock();\n String full = block.getRegistryName().getNamespace()+\":\"+block.getRegistryName().getPath();\n return WoodenCogCommonConfigs.WEAR_BLACKLIST.get().contains(full);\n }", "score": 0.8505130410194397 }, { "filename": "src/main/java/net/chauvedev/woodencog/core/MachineCapacityStorage.java", "retrieved_chunk": " public MachineCapacityEntry getCapacity(Block block){\n String full = block.getRegistryName().getNamespace()+\":\"+block.getRegistryName().getPath();\n try {\n CustomBlockConfig.BlockInformation info = CustomBlockConfig.registeredBlocks.get(full);\n return MachineCapacityEntry.createEntryBlock(false,\n info.durability,\n info.chance\n );\n } catch (NullPointerException e) {\n return MachineCapacityEntry.createEntryBlock(false,", "score": 0.8472939729690552 }, { "filename": "src/main/java/net/chauvedev/woodencog/WoodenCogEvents.java", "retrieved_chunk": " Block block = ((BlockItem) event.getObject().getItem()).getBlock();\n if (block instanceof KineticBlock){\n event.addCapability(MachineCapacityProvider.MACHINE_CAPACITY_KEY, new MachineCapacityProvider());\n }\n }\n }catch (Error error){\n WoodenCog.LOGGER.debug(\"-------ERROR---\");\n WoodenCog.LOGGER.error(\"Error found\",error);\n }\n }", "score": 0.8469783067703247 }, { "filename": "src/main/java/net/chauvedev/woodencog/WoodenCogEvents.java", "retrieved_chunk": " @Mod.EventBusSubscriber(modid = WoodenCog.MOD_ID)\n public static class ForgeEvents {\n @SubscribeEvent\n public static void onAttachCapabilitiesBlockEntity(AttachCapabilitiesEvent<BlockEntity> event) {\n if (! event.getCapabilities().containsKey(MachineCapacityProvider.MACHINE_CAPACITY_KEY)){\n try {\n if(event.getObject() instanceof KineticBlockEntity) {\n event.addCapability(MachineCapacityProvider.MACHINE_CAPACITY_KEY, new MachineCapacityProvider());\n }\n }catch (Error error){", "score": 0.829049825668335 }, { "filename": "src/main/java/net/chauvedev/woodencog/mixin/MixinKineticBlock.java", "retrieved_chunk": "package net.chauvedev.woodencog.mixin;\nimport com.simibubi.create.content.kinetics.base.IRotate;\nimport com.simibubi.create.content.kinetics.base.KineticBlock;\nimport com.simibubi.create.content.kinetics.base.KineticBlockEntity;\nimport net.chauvedev.woodencog.capability.MachineCapacity;\nimport net.chauvedev.woodencog.capability.MachineCapacityProvider;\nimport net.dries007.tfc.common.TFCTags;\nimport net.minecraft.client.resources.sounds.Sound;\nimport net.minecraft.core.BlockPos;\nimport net.minecraft.server.level.ServerPlayer;", "score": 0.8261941075325012 } ]
java
MachineCapacityEntry config = MachineCapacityStorage.getInstance().getCapacity(block.getBlockState().getBlock());
package net.chauvedev.woodencog.mixin; import com.simibubi.create.content.kinetics.KineticNetwork; import com.simibubi.create.content.kinetics.base.KineticBlockEntity; import com.simibubi.create.content.kinetics.crusher.CrushingWheelBlockEntity; import com.simibubi.create.content.kinetics.transmission.ClutchBlockEntity; import net.chauvedev.woodencog.WoodenCog; import net.chauvedev.woodencog.capability.MachineCapacity; import net.chauvedev.woodencog.capability.MachineCapacityProvider; import net.chauvedev.woodencog.core.MachineCapacityEntry; import net.chauvedev.woodencog.core.MachineCapacityStorage; import net.minecraft.core.BlockPos; import net.minecraft.core.particles.ParticleTypes; import net.minecraft.nbt.CompoundTag; import net.minecraft.sounds.SoundEvent; import net.minecraft.sounds.SoundEvents; import net.minecraft.sounds.SoundSource; import net.minecraft.world.level.Level; import net.minecraft.world.level.block.Block; import net.minecraft.world.level.block.Blocks; import net.minecraft.world.level.block.entity.BlockEntity; import net.minecraft.world.level.block.entity.BlockEntityType; import net.minecraft.world.level.block.state.BlockState; import org.spongepowered.asm.mixin.Mixin; import org.spongepowered.asm.mixin.Shadow; import org.spongepowered.asm.mixin.injection.At; import org.spongepowered.asm.mixin.injection.Inject; import org.spongepowered.asm.mixin.injection.callback.CallbackInfo; import java.awt.*; import java.util.Objects; @Mixin(value = KineticBlockEntity.class, remap = false) public abstract class MixinKineticBlockEntity{ private boolean lifeBlockBlocked = false; private float lifeBlock = 20*10; @Shadow public abstract KineticNetwork getOrCreateNetwork(); @Shadow protected float speed; @Shadow protected float stress; @Shadow protected boolean overStressed; @Shadow public abstract void setSpeed(float speed); @Shadow protected float capacity; @Shadow public abstract float calculateAddedStressCapacity(); @Shadow public abstract float calculateStressApplied(); @Shadow protected float lastStressApplied; @Shadow protected float lastCapacityProvided; @Shadow public abstract void detachKinetics(); @Shadow public abstract float getSpeed(); boolean destroyed = false; @Inject( method = {"tick"}, at = {@At("HEAD")}, remap = false) public void tick(CallbackInfo ci){ KineticBlockEntity block = (KineticBlockEntity)((Object)this) ; try{ this.tickDamagedTick(block); }catch (Error error){ WoodenCog.LOGGER.debug("Error pendant le system tick"); error.printStackTrace(); } } public void tickDamagedTick(KineticBlockEntity block){ if (!MachineCapacityStorage.getInstance().active){ return; } if (MachineCapacityStorage.getInstance().isBlackListBlock(block)){ return; } MachineCapacityEntry config = MachineCapacityStorage.getInstance().getCapacity(block.getBlockState().getBlock()); MachineCapacity capacity = block.getCapability(MachineCapacityProvider.MACHINE_CAPACITY).resolve().get(); int chance = block.getLevel().random.nextInt(0,101); float left = config.durabilityMax - capacity.getDurability(); if (chance > (100-config.damageChance) && left > 0 && block.getSpeed() > 0){ int damage_for_speed = (int) getSpeed()/10; capacity.setDurability
(capacity.getDurability()+damage_for_speed);
} left = config.durabilityMax - capacity.getDurability(); if(left >= 10 && left < config.durabilityMax/2) { block.getLevel().addParticle(ParticleTypes.SMOKE,false,block.getBlockPos().getX()+0.5f,block.getBlockPos().getY()+0.5f,block.getBlockPos().getZ()+0.5f,0,0.01,0); } if (left<10){ if (getOrCreateNetwork()!=null){ if (!destroyed || !capacity.isDestroyed() || this.stress != Integer.MAX_VALUE){ destroyed = true; getOrCreateNetwork().updateCapacityFor(block,0); getOrCreateNetwork().updateStressFor(block, Integer.MAX_VALUE); getOrCreateNetwork().updateNetwork(); getOrCreateNetwork().sync(); capacity.setDestroyed(true); } } block.getLevel().addParticle(ParticleTypes.CAMPFIRE_SIGNAL_SMOKE,false,block.getBlockPos().getX()+0.5f,block.getBlockPos().getY()+0.5f,block.getBlockPos().getZ()+0.5f,0,0.01,0); } } }
src/main/java/net/chauvedev/woodencog/mixin/MixinKineticBlockEntity.java
Sparks-and-Rotation-woodencog-bb3d3ef
[ { "filename": "src/main/java/net/chauvedev/woodencog/core/MachineCapacityStorage.java", "retrieved_chunk": " public MachineCapacityEntry getCapacity(Block block){\n String full = block.getRegistryName().getNamespace()+\":\"+block.getRegistryName().getPath();\n try {\n CustomBlockConfig.BlockInformation info = CustomBlockConfig.registeredBlocks.get(full);\n return MachineCapacityEntry.createEntryBlock(false,\n info.durability,\n info.chance\n );\n } catch (NullPointerException e) {\n return MachineCapacityEntry.createEntryBlock(false,", "score": 0.8673301339149475 }, { "filename": "src/main/java/net/chauvedev/woodencog/mixin/MixinKineticBlock.java", "retrieved_chunk": " } else {\n capacity.setDurability(0);\n int durability_to_remove = (int)(capacity.getDurability() / 10) + 1;\n itemInHand.hurtAndBreak(durability_to_remove, pPlayer, (player -> player.broadcastBreakEvent(player.getUsedItemHand())));\n pLevel.playLocalSound(\n pPos.getX(),\n pPos.getY(),\n pPos.getZ(),\n SoundEvents.ANVIL_USE,\n SoundSource.BLOCKS,", "score": 0.8569043874740601 }, { "filename": "src/main/java/net/chauvedev/woodencog/WoodenCogEvents.java", "retrieved_chunk": " Block block = ((BlockItem) event.getObject().getItem()).getBlock();\n if (block instanceof KineticBlock){\n event.addCapability(MachineCapacityProvider.MACHINE_CAPACITY_KEY, new MachineCapacityProvider());\n }\n }\n }catch (Error error){\n WoodenCog.LOGGER.debug(\"-------ERROR---\");\n WoodenCog.LOGGER.error(\"Error found\",error);\n }\n }", "score": 0.8329797983169556 }, { "filename": "src/main/java/net/chauvedev/woodencog/capability/MachineCapacity.java", "retrieved_chunk": "package net.chauvedev.woodencog.capability;\nimport net.minecraft.nbt.CompoundTag;\nimport org.checkerframework.checker.units.qual.C;\npublic class MachineCapacity {\n private float machine_capacity = 0;\n private boolean destroyed = false;\n public float getDurability(){\n return this.machine_capacity;\n }\n public void setDurability(float var1){", "score": 0.8319416642189026 }, { "filename": "src/main/java/net/chauvedev/woodencog/config/WoodenCogCommonConfigs.java", "retrieved_chunk": " .define(\"durability\", 144000);\n DEFAULT_DAMAGE_CHANCE = BUILDER\n .comment(\"chance of machine getting damage (from 0 to 100 number over 100 are the same as 100)\")\n .define(\"chance\", 10);\n WEAR_BLACKLIST = BUILDER\n .comment(\"This list contains block that should not damage over time\")\n .defineList(\"blacklist\", List.of(\n \"create:shaft\",\n \"create:cogwheel\",\n \"create:large_cogwheel\",", "score": 0.8292829990386963 } ]
java
(capacity.getDurability()+damage_for_speed);