Development of Big Data Module--Statistical Analysis

After the data warehouse is built, users can write Hive SQL statements to access them and analyze the data. In actual production, which statistical indicators are needed are usually proposed by the relevant departments of data demand, and new statistical requirements will continue to emerge. The following are some typical indicators in website ...

Added by davey10101 on Fri, 23 Aug 2019 06:58:37 +0300

Arrays, mappings, tuples, collections of spark notes

1.1. Array 1.1.1. Fixed-length and variable-length arrays (1) Definition format of fixed-length array: val arr=new ArrayT (2) Variable-length array definition format: val arr = ArrayBuffer[T]() Note the need for a guide: import scala. collection. mutable. Array Buffer package cn.itcast.scala import scala.collection.mutable.ArrayBuffer object ...

Added by Avi on Tue, 13 Aug 2019 10:14:47 +0300

Data mining: MATLAB decision tree (using wine data of UCI data set), and drawing, drawing at the end

There is nothing to do during the summer vacation, thinking about data mining, this blog is also written in the case of my first study of MATLAB (code can be placed directly in a file). On the one hand, I want to deepen my understanding, on the other hand, I hope that I can give a reference to the reade ...

Added by Afrojojo on Sat, 10 Aug 2019 11:42:24 +0300

Zero-Foundation Learning Java - - - day4 - - - Process Control Architecture

1. Sequential structure Code is executed from top to bottom 2. Selection structure Also known as branch structure, it will choose different code execution according to the results of execution, there are the following two forms: if statement switch statement 2.1 if statement 2.1.1 The first form of an if statement Format: If (relatio ...

Added by l00ph0le on Sun, 04 Aug 2019 16:43:39 +0300

Filling mechanism of default parameters in kubelet service startup process

This article traces how the default values of parameters are injected during kubelet startup. We know that in order to start the kubelet service, we first need to construct the configuration object of kubelet, that is, the kubeletconfig.KubeletConfiguration structure. // NewKubeletCommand creates a *cob ...

Added by cahva on Wed, 31 Jul 2019 16:29:39 +0300

My docker essay 14: An example of MySQL operation

This article uses docker to run MySQL and do simple operations. This article uses MySQL as the official mirror, version 5.7. Download the mirror command: docker pull mysql:5.7 Run the container with the following command and set the root password to 123456. run --name hi-mysql -e MYSQL_ROOT_PASSWORD=1 ...

Added by widox on Wed, 31 Jul 2019 01:59:01 +0300

Java Implementation of Kafka Tools and the like in 59-Dawn Big Data Classic Big Data Written Test Series

Tool Class Meaning Description kafka is one of the most typical and commonly used MQ technologies. It has gradually developed into a middleware for decoupling components of producers and consumers. code implementation Producer Tools package com.tl.job002.kafka; import java.util.Properties; import o ...

Added by olko on Tue, 30 Jul 2019 14:05:03 +0300

My docker essays 15: MySQL automatically creates databases at startup

Background and Analysis When the MySQL container starts up, it automatically creates some necessary databases, such as MySQL, which is the official default. However, in practice, we also need to let MySQL create our own custom database automatically. This paper explores the application occasion. The ...

Added by rekha on Mon, 29 Jul 2019 17:14:18 +0300

python Stock Market Data Exploration

Preface Although the data of financial management applications such as Tonghuashun is good enough, there is still an impulse to customize the data. Naturally, the data will not be much better than the former, but it is good to customize it according to your own ideas. target Data is obtained through a free data interface, and the historical tr ...

Added by mahenderp on Sat, 27 Jul 2019 17:14:49 +0300

Preparations for Hadoop Cluster Building-01

The whole process of building hadoop cluster includes preparation in advance Install zookeeper and configure the environment Compile, install and start hadoop Install HDFS to manage namenode and dataname to manage cluster hard disk resources Install and start yarn to build MapReduce to manage cpu and memory resources 01 Preparation ...

Added by gasxtreme on Sun, 21 Jul 2019 14:34:55 +0300