kafka-rest:怎么愉快的build?
愉快的build该项目吧
git clone https://github.com/confluentinc/kafka-rest
mvn clean install -Dmaven.test.skip=true
出先如下类似错误:
[ERROR] Failed to execute goal on project kafka-rest: Could not resolve dependencies for project io.confluent:kafka-rest:jar:3.1.-SNAPSHOT: The following artifacts could not be resolved: io.confluent:rest-utils:jar:3.1.-SNAPSHOT, org.apache.kafka:kafka_2.:jar:0.10.1.0-SNAPSHOT, io.confluent:kafka-avro-serializer:jar:3.1.-SNAPSHOT, io.confluent:kafka-json-serializer:jar:3.1.-SNAPSHOT, io.confluent:rest-utils-test:jar:3.1.-SNAPSHOT, org.apache.kafka:kafka_2.:jar:test:0.10.1.0-SNAPSHOT, org.apache.kafka:kafka-clients:jar:test:0.10.1.0-SNAPSHOT, io.confluent:kafka-schema-registry:jar:3.1.-SNAPSHOT: Could not find artifact io.confluent:rest-utils:jar:3.1.-SNAPSHOT in confluent (http://packages.confluent.io/maven/) -> [Help 1]
翻了一下该项目wiki:
https://github.com/confluentinc/kafka-rest/wiki/FAQ 描述如下:
When I try to compile I get an error about a missing SNAPSHOT dependency.
The error looks something like this:
[ERROR] Failed to execute goal on project kafka-rest: Could not resolve dependencies for project io.confluent:kafka-rest:jar:3.1.0-SNAPSHOT: The following artifacts could not be resolved: io.confluent:rest-utils:jar:3.1.0-SNAPSHOT, org.apache.kafka:kafka_2.11:jar:0.10.1.0-SNAPSHOT, io.confluent:kafka-avro-serializer:jar:3.1.0-SNAPSHOT, io.confluent:kafka-json-serializer:jar:3.1.0-SNAPSHOT, io.confluent:rest-utils-test:jar:3.1.0-SNAPSHOT, org.apache.kafka:kafka_2.11:jar:test:0.10.1.0-SNAPSHOT, org.apache.kafka:kafka-clients:jar:test:0.10.1.0-SNAPSHOT, io.confluent:kafka-schema-registry:jar:3.1.0-SNAPSHOT: Could not find artifact io.confluent:rest-utils:jar:3.1.0-SNAPSHOT in confluent (http://packages.confluent.io/maven/) -> [Help 1]
During development of new versions of the REST Proxy we sometimes use a SNAPSHOT versions of dependencies in order to build and test against new features. If you want to build a development version, you may need to build and install these dependencies to your local Maven repository in order to build the connector:
- Kafka - clone https://github.com/apache/kafka/ and build with
./gradlew installAll - rest-utils - clone https://github.com/confluentinc/rest-utils and build with
mvn clean install - schema-registry - clone https://github.com/confluentinc/schema-registry and build with
mvn clean install
按照上述说法,一通搞,还是不行
翻了一下该项目Issues 描述如下:
https://github.com/confluentinc/common/issues/98
@ewencp thanks for the summary, that finally also worked for me. I used the following steps (a bit in more detail) to build it.
Actually our company network (proxy and maven repo) was a problem since could either resolve dependencies or run the integration tests. Therefore, I had to skip the tests. git clone --branch 1.1. https://github.com/apache/kafka.git
gradle wrapper
./gradlew installAll git clone https://github.com/confluentinc/common.git
git checkout origin/4.1.x
mvn clean install -Dmaven.test.skip=true git clone https://github.com/confluentinc/rest-utils.git
git checkout origin/4.1.x
mvn clean install -Dmaven.test.skip=true git clone https://github.com/confluentinc/schema-registry.git
git checkout origin/4.1.x
mvn clean install -Dmaven.test.skip=true git clone https://github.com/confluentinc/kafka-rest.git
git checkout origin/4.1.x
mvn clean install -Dmaven.test.skip=true
按照上述顺序,终于愉快的build了。。。
当然,如果没打算基于源码最改动,可以按照该项目Readme建议的,直接使用prebuilt版本
Installation
You can download prebuilt versions of the Kafka REST Proxy as part of the Confluent Platform.
You can read our full installation instructions and the complete documentation
To install from source, follow the instructions in the Development section below.
kafka-rest:怎么愉快的build?的更多相关文章
- kafka-rest:A Comprehensive, Open Source REST Proxy for Kafka
Ewen Cheslack-Postava March 25, 2015 时间有点久,但讲的还是很清楚的 As part of Confluent Platform 1.0 released ab ...
- Java连接kafka
1.maven依赖: <project xmlns="http://maven.apache.org/POM/4.0.0" xmlns:xsi="http://ww ...
- 056 Java搭建kafka环境
1.使用Java项目搭建 2.新目录 3.添加项目支持 4.添加mavem与scala 5.修改pom <?xml version="1.0" encoding=" ...
- 搭建kafka源码开发环境时使用"gradle idea"命令构建源码失败
我的环境: JDK: 1.8.0_131 Gradle: Gradle 3.1 Kafka源码包: kafka-0.10.0.1-src.tgz Zookeeper安装包: zookeeper-3.4 ...
- Openshift部署Zookeeper和Kafka
部署Zookeeper github网址 https://github.com/ericnie2015/zookeeper-k8s-openshift 1.在openshift目录中,首先构建imag ...
- kfka学习笔记一:使用Python操作Kafka
1.准备工作 使用python操作kafka目前比较常用的库是kafka-python库,但是在安装这个库的时候需要依赖setuptools库和six库,下面就要分别来下载这几个库 https://p ...
- 使用sbt构建spark 程序
今日在学习scala和spark相关的知识.之前在eclipse下编写了wordcount程序.但是关于导出jar包这块还是很困惑.于是学习sbt构建scala. 关于sbt的介绍网上有很多的资料,这 ...
- Flink - FlinkKafkaConsumer08
先看 AbstractFetcher 这个可以理解就是,consumer中具体去kafka读数据的线程,一个fetcher可以同时读多个partitions的数据来看看 /** * Base cl ...
- 大数据系列之kafka-java实现
Java源码GitBub地址: https://github.com/fzmeng/kafka-demo 关于kafka安装步骤可见文章 http://www.cnblogs.com/cnmeng ...
随机推荐
- 为了学好Java,我尝试了这 6 个方法
阅读本文大概需要 5 分钟. 教练,我想学Java! 怎么学Java,一个简单的命题,我自己也折腾了好几年,现在虽不能说是Java高手,但也算是小有所成,至少还不至于搞不懂一些基本概念和技术原理. 从 ...
- [Leetcode]538. Convert BST to Greater Tree
Given a Binary Search Tree (BST), convert it to a Greater Tree such that every key of the original B ...
- 面试 Linux 运维工作至少需要知道哪些知识?
前言 我们已经发过不少 Linux 面试题,但是单独的面试题总感觉会过于零碎,没有体系化内容给人的帮助大. 知乎上有这样一个问题:一个新手面试 Linux 运维工作至少需要知道哪些知识?其中有一个答案 ...
- Chapter 5 Blood Type——16
"Okay, guys, I want you all to take one piece from each box," he said as he produced a pai ...
- 在AspNetCore中使用极验做行为认证
先上效果图 极验的流程 极验官方文档地址 https://docs.geetest.com/install/deploy/server/csharp 简单说明一下极验的验证流程 引用官方的图片 向服务 ...
- logback配置信息
<?xml version="1.0" encoding="UTF-8"?> <configuration scan="true&q ...
- 为容器化的 Go 程序搭建 CI
本文介绍如何使用 Jenkins 的声明式 pipeline 为一个简单的 Golang web 应用搭建 CI 环境.如果你还不太了解 Jenkins 及其声明式 pipeline,请先参考笔者的 ...
- Java线程Thread的状态解析以及状态转换分析 多线程中篇(七)
线程与操作系统中线程(进程)的概念同根同源,尽管千差万别. 操作系统中有状态以及状态的切换,Java线程中照样也有. State 在Thread类中有内部类 枚举State,用于抽象描述Java线程的 ...
- 痞子衡嵌入式:第一本Git命令教程(7.1)- 清理之缓存(stash)
今天是Git系列课程第七课,上一课我们学会了查看Git本地历史提交,今天痞子衡要讲的是Git仓库的清理操作,一共4个命令,都是日常开发中非常实用的命令,掌握这4个命令,会让你有一种玩弄Git仓库于股掌 ...
- k8s运行容器之deployment(三)--技术流ken
deployment 我们已经知道k8s是通过各种controller来管理pod的生命周期.为了满足不同业务场景,k8s开发了Deployment.ReplicaSet.DaemonSet.Stat ...