网站建设资讯

NEWS

网站建设资讯

flink从mysql读取数据放入kafka中用于搜索全量

接着上一篇,将MySQL的数据导入kafka中

临夏州网站建设公司成都创新互联,临夏州网站设计制作,有大型网站制作公司丰富经验。已为临夏州上1000家提供企业网站建设服务。企业网站搭建\成都外贸网站建设要多少钱,请找那个售后服务好的临夏州做网站的公司定做!

public static void main(String[] arg) throws Exception {

        TypeInformation[] fieldTypes = new TypeInformation[] { BasicTypeInfo.STRING_TYPE_INFO, BasicTypeInfo.STRING_TYPE_INFO };
        String[] fieldNames = new String[] { "name", "address" };

        RowTypeInfo rowTypeInfo = new RowTypeInfo(fieldTypes, fieldNames);
        JDBCInputFormat jdbcInputFormat = JDBCInputFormat.buildJDBCInputFormat().setDrivername("com.mysql.jdbc.Driver")
                .setDBUrl("jdbc:mysql://ip:3306/tablespace?characterEncoding=utf8")
                .setUsername("user").setPassword("root")
                .setQuery("select LOGIC_CODE, SHARE_LOG_CODE from table").setRowTypeInfo(rowTypeInfo).finish();

        final ExecutionEnvironment env = ExecutionEnvironment.getExecutionEnvironment();
        DataSource s = env.createInput(jdbcInputFormat);

        BatchTableEnvironment tableEnv = new BatchTableEnvironment(env, TableConfig.DEFAULT());

        tableEnv.registerDataSet("t2", s);

        Table tapiResult = tableEnv.scan("t2");
        System.out.println("schema is:");
        tapiResult.printSchema();

        Table query = tableEnv.sqlQuery("select name, address from t2");
        DataSet ds=  tableEnv.toDataSet(query, Result.class);

        DataSet temp=ds.map(new MapFunction() {
            @Override
            public String map(Result result) throws Exception {
                String name = result.name;
                String value = result.address;
                return name+":->:"+value;
            }
        });

        logger.info("read db end"); 

        KafkaOutputFormat kafkaOutput = KafkaOutputFormat.buildKafkaOutputFormat()
                .setBootstrapServers("ip:9092").setTopic("search_test_whk").setAcks("all").setBatchSize("1000")
                .setBufferMemory("100000").setLingerMs("1").setRetries("2").finish();

        temp.output(kafkaOutput);

        logger.info("write kafka end");

        env.execute("Flink add data source");

    }

文章标题:flink从mysql读取数据放入kafka中用于搜索全量
链接分享:http://cdweb.net/article/jgchdp.html