Development of Big Data Module--Statistical Analysis
After the data warehouse is built, users can write Hive SQL statements to access them and analyze the data.
In actual production, which statistical indicators are needed are usually proposed by the relevant departments of data demand, and new statistical requirements will continue to emerge. The following are some typical indicators in website ...
Added by davey10101 on Fri, 23 Aug 2019 06:58:37 +0300
Arrays, mappings, tuples, collections of spark notes
1.1. Array 1.1.1. Fixed-length and variable-length arrays
(1) Definition format of fixed-length array:
val arr=new ArrayT
(2) Variable-length array definition format:
val arr = ArrayBuffer[T]()
Note the need for a guide: import scala. collection. mutable. Array Buffer
package cn.itcast.scala
import scala.collection.mutable.ArrayBuffer
object ...
Added by Avi on Tue, 13 Aug 2019 10:14:47 +0300
Data mining: MATLAB decision tree (using wine data of UCI data set), and drawing, drawing at the end
There is nothing to do during the summer vacation, thinking about data mining, this blog is also written in the case of my first study of MATLAB (code can be placed directly in a file).
On the one hand, I want to deepen my understanding, on the other hand, I hope that I can give a reference to the reade ...
Added by Afrojojo on Sat, 10 Aug 2019 11:42:24 +0300
Zero-Foundation Learning Java - - - day4 - - - Process Control Architecture
1. Sequential structure
Code is executed from top to bottom
2. Selection structure
Also known as branch structure, it will choose different code execution according to the results of execution, there are the following two forms:
if statement
switch statement
2.1 if statement
2.1.1 The first form of an if statement
Format:
If (relatio ...
Added by l00ph0le on Sun, 04 Aug 2019 16:43:39 +0300
Filling mechanism of default parameters in kubelet service startup process
This article traces how the default values of parameters are injected during kubelet startup.
We know that in order to start the kubelet service, we first need to construct the configuration object of kubelet, that is, the kubeletconfig.KubeletConfiguration structure.
// NewKubeletCommand creates a *cob ...
Added by cahva on Wed, 31 Jul 2019 16:29:39 +0300
My docker essay 14: An example of MySQL operation
This article uses docker to run MySQL and do simple operations.
This article uses MySQL as the official mirror, version 5.7. Download the mirror command:
docker pull mysql:5.7
Run the container with the following command and set the root password to 123456.
run --name hi-mysql -e MYSQL_ROOT_PASSWORD=1 ...
Added by widox on Wed, 31 Jul 2019 01:59:01 +0300
Java Implementation of Kafka Tools and the like in 59-Dawn Big Data Classic Big Data Written Test Series
Tool Class Meaning Description
kafka is one of the most typical and commonly used MQ technologies. It has gradually developed into a middleware for decoupling components of producers and consumers.
code implementation
Producer Tools
package com.tl.job002.kafka;
import java.util.Properties;
import o ...
Added by olko on Tue, 30 Jul 2019 14:05:03 +0300
My docker essays 15: MySQL automatically creates databases at startup
Background and Analysis
When the MySQL container starts up, it automatically creates some necessary databases, such as MySQL, which is the official default. However, in practice, we also need to let MySQL create our own custom database automatically. This paper explores the application occasion.
The ...
Added by rekha on Mon, 29 Jul 2019 17:14:18 +0300
python Stock Market Data Exploration
Preface
Although the data of financial management applications such as Tonghuashun is good enough, there is still an impulse to customize the data. Naturally, the data will not be much better than the former, but it is good to customize it according to your own ideas.
target
Data is obtained through a free data interface, and the historical tr ...
Added by mahenderp on Sat, 27 Jul 2019 17:14:49 +0300
Preparations for Hadoop Cluster Building-01
The whole process of building hadoop cluster includes
preparation in advance
Install zookeeper and configure the environment
Compile, install and start hadoop
Install HDFS to manage namenode and dataname to manage cluster hard disk resources
Install and start yarn to build MapReduce to manage cpu and memory resources
01 Preparation ...
Added by gasxtreme on Sun, 21 Jul 2019 14:34:55 +0300