|
24 | 24 | import java.util.HashMap;
|
25 | 25 | import java.util.List;
|
26 | 26 | import java.util.Map;
|
| 27 | +import java.util.Optional; |
27 | 28 | import java.util.concurrent.ExecutionException;
|
28 | 29 | import java.util.concurrent.TimeUnit;
|
29 | 30 | import java.util.concurrent.TimeoutException;
|
30 | 31 | import java.util.stream.Collectors;
|
31 | 32 |
|
32 | 33 | import org.apache.commons.logging.LogFactory;
|
33 | 34 | import org.apache.kafka.clients.admin.AdminClient;
|
| 35 | +import org.apache.kafka.clients.admin.AlterConfigOp; |
| 36 | +import org.apache.kafka.clients.admin.AlterConfigOp.OpType; |
| 37 | +import org.apache.kafka.clients.admin.AlterConfigsResult; |
| 38 | +import org.apache.kafka.clients.admin.Config; |
| 39 | +import org.apache.kafka.clients.admin.ConfigEntry; |
34 | 40 | import org.apache.kafka.clients.admin.CreatePartitionsResult;
|
35 | 41 | import org.apache.kafka.clients.admin.CreateTopicsResult;
|
| 42 | +import org.apache.kafka.clients.admin.DescribeConfigsResult; |
36 | 43 | import org.apache.kafka.clients.admin.DescribeTopicsResult;
|
37 | 44 | import org.apache.kafka.clients.admin.NewPartitions;
|
38 | 45 | import org.apache.kafka.clients.admin.NewTopic;
|
39 | 46 | import org.apache.kafka.clients.admin.TopicDescription;
|
| 47 | +import org.apache.kafka.common.config.ConfigResource; |
| 48 | +import org.apache.kafka.common.config.ConfigResource.Type; |
40 | 49 | import org.apache.kafka.common.errors.InvalidPartitionsException;
|
41 | 50 | import org.apache.kafka.common.errors.TopicExistsException;
|
42 | 51 | import org.apache.kafka.common.errors.UnsupportedVersionException;
|
@@ -237,13 +246,99 @@ private void addOrModifyTopicsIfNeeded(AdminClient adminClient, Collection<NewTo
|
237 | 246 | .map(NewTopic::name)
|
238 | 247 | .collect(Collectors.toList()));
|
239 | 248 | List<NewTopic> topicsToAdd = new ArrayList<>();
|
240 |
| - Map<String, NewPartitions> topicsToModify = checkPartitions(topicNameToTopic, topicInfo, topicsToAdd); |
| 249 | + Map<String, NewPartitions> topicsWithPartitionMismatches = |
| 250 | + checkPartitions(topicNameToTopic, topicInfo, topicsToAdd); |
241 | 251 | if (topicsToAdd.size() > 0) {
|
242 | 252 | addTopics(adminClient, topicsToAdd);
|
243 | 253 | }
|
244 |
| - if (topicsToModify.size() > 0) { |
245 |
| - modifyTopics(adminClient, topicsToModify); |
| 254 | + if (topicsWithPartitionMismatches.size() > 0) { |
| 255 | + createMissingPartitions(adminClient, topicsWithPartitionMismatches); |
246 | 256 | }
|
| 257 | + Map<ConfigResource, List<ConfigEntry>> mismatchingConfigs = |
| 258 | + checkTopicsForConfigMismatches(adminClient, topics); |
| 259 | + if (!mismatchingConfigs.isEmpty()) { |
| 260 | + adjustConfigMismatches(adminClient, topics, mismatchingConfigs); |
| 261 | + } |
| 262 | + } |
| 263 | + } |
| 264 | + |
| 265 | + private Map<ConfigResource, List<ConfigEntry>> checkTopicsForConfigMismatches( |
| 266 | + AdminClient adminClient, Collection<NewTopic> topics) { |
| 267 | + List<ConfigResource> configResources = topics.stream() |
| 268 | + .map(topic -> new ConfigResource(Type.TOPIC, topic.name())) |
| 269 | + .collect(Collectors.toList()); |
| 270 | + |
| 271 | + DescribeConfigsResult describeConfigsResult = adminClient.describeConfigs(configResources); |
| 272 | + try { |
| 273 | + Map<ConfigResource, Config> topicsConfig = describeConfigsResult.all() |
| 274 | + .get(this.operationTimeout, TimeUnit.SECONDS); |
| 275 | + |
| 276 | + Map<ConfigResource, List<ConfigEntry>> configMismatches = new HashMap<>(); |
| 277 | + for (Map.Entry<ConfigResource, Config> topicConfig : topicsConfig.entrySet()) { |
| 278 | + Optional<NewTopic> topicOptional = topics.stream() |
| 279 | + .filter(p -> p.name().equals(topicConfig.getKey().name())) |
| 280 | + .findFirst(); |
| 281 | + |
| 282 | + List<ConfigEntry> configMismatchesEntries = new ArrayList<>(); |
| 283 | + if (topicOptional.isPresent() && topicOptional.get().configs() != null) { |
| 284 | + for (Map.Entry<String, String> desiredConfigParameter : topicOptional.get().configs().entrySet()) { |
| 285 | + ConfigEntry actualConfigParameter = topicConfig.getValue().get(desiredConfigParameter.getKey()); |
| 286 | + if (!actualConfigParameter.value().equals(desiredConfigParameter.getValue())) { |
| 287 | + configMismatchesEntries.add(actualConfigParameter); |
| 288 | + } |
| 289 | + |
| 290 | + if (configMismatchesEntries.size() > 0) { |
| 291 | + configMismatches.put(topicConfig.getKey(), configMismatchesEntries); |
| 292 | + } |
| 293 | + } |
| 294 | + } |
| 295 | + } |
| 296 | + return configMismatches; |
| 297 | + } |
| 298 | + catch (InterruptedException ie) { |
| 299 | + Thread.currentThread().interrupt(); |
| 300 | + throw new KafkaException("Interrupted while getting topic descriptions", ie); |
| 301 | + } |
| 302 | + catch (ExecutionException | TimeoutException ex) { |
| 303 | + throw new KafkaException("Failed to obtain topic descriptions", ex); |
| 304 | + } |
| 305 | + } |
| 306 | + |
| 307 | + private void adjustConfigMismatches(AdminClient adminClient, Collection<NewTopic> topics, |
| 308 | + Map<ConfigResource, List<ConfigEntry>> mismatchingConfigs) { |
| 309 | + for (Map.Entry<ConfigResource, List<ConfigEntry>> mismatchingConfigsOfTopic : mismatchingConfigs.entrySet()) { |
| 310 | + ConfigResource topicConfigResource = mismatchingConfigsOfTopic.getKey(); |
| 311 | + |
| 312 | + Optional<NewTopic> topicOptional = topics.stream().filter(p -> p.name().equals(topicConfigResource.name())) |
| 313 | + .findFirst(); |
| 314 | + if (topicOptional.isPresent()) { |
| 315 | + for (ConfigEntry mismatchConfigEntry : mismatchingConfigsOfTopic.getValue()) { |
| 316 | + List<AlterConfigOp> alterConfigOperations = new ArrayList<>(); |
| 317 | + Map<String, String> desiredConfigs = topicOptional.get().configs(); |
| 318 | + if (desiredConfigs.get(mismatchConfigEntry.name()) != null) { |
| 319 | + alterConfigOperations.add( |
| 320 | + new AlterConfigOp( |
| 321 | + new ConfigEntry(mismatchConfigEntry.name(), |
| 322 | + desiredConfigs.get(mismatchConfigEntry.name())), |
| 323 | + OpType.SET)); |
| 324 | + } |
| 325 | + if (alterConfigOperations.size() > 0) { |
| 326 | + try { |
| 327 | + AlterConfigsResult alterConfigsResult = adminClient |
| 328 | + .incrementalAlterConfigs(Map.of(topicConfigResource, alterConfigOperations)); |
| 329 | + alterConfigsResult.all().get(this.operationTimeout, TimeUnit.SECONDS); |
| 330 | + } |
| 331 | + catch (InterruptedException ie) { |
| 332 | + Thread.currentThread().interrupt(); |
| 333 | + throw new KafkaException("Interrupted while getting topic descriptions", ie); |
| 334 | + } |
| 335 | + catch (ExecutionException | TimeoutException ex) { |
| 336 | + throw new KafkaException("Failed to obtain topic descriptions", ex); |
| 337 | + } |
| 338 | + } |
| 339 | + } |
| 340 | + } |
| 341 | + |
247 | 342 | }
|
248 | 343 | }
|
249 | 344 |
|
@@ -304,7 +399,7 @@ private void addTopics(AdminClient adminClient, List<NewTopic> topicsToAdd) {
|
304 | 399 | }
|
305 | 400 | }
|
306 | 401 |
|
307 |
| - private void modifyTopics(AdminClient adminClient, Map<String, NewPartitions> topicsToModify) { |
| 402 | + private void createMissingPartitions(AdminClient adminClient, Map<String, NewPartitions> topicsToModify) { |
308 | 403 | CreatePartitionsResult partitionsResult = adminClient.createPartitions(topicsToModify);
|
309 | 404 | try {
|
310 | 405 | partitionsResult.all().get(this.operationTimeout, TimeUnit.SECONDS);
|
|
0 commit comments