Building Blog Websites from Zero Based on SpringBook - Adding Column Creation, Modification and Deletation Functions

Watch Blog is a function that supports the creation of columns, that is, a series of related articles can be archived into columns, to facilitate user management and access to articles. Here we mainly explain the creation, modification and deletion of columns. As for the columns, there are other functions involved, such as attention columns, wh ...

Added by BillyBoB on Tue, 30 Jul 2019 08:43:50 +0300

Distributed Cluster Sharing Session of SpringBook Development Case

Preface In distributed systems, in order to improve system performance, individual projects are usually split into multiple function-based micro-services. If conditions permit, single micro-services may be scaled horizontally to ensure high availability of services. So the question is, what kind of problems will we encounter if we use the trad ...

Added by pacome on Tue, 30 Jul 2019 08:30:56 +0300

Monitor

Listener introduce The listener comes from the servlet specification The listener is dedicated to monitoring [changes in domain object lifecycle] and [changes in domain object shared data] The implementation class of listener interface must be implemented by the developer As mentioned above, listeners are used to monitor changes in the life c ...

Added by mizkit73 on Tue, 30 Jul 2019 07:24:25 +0300

KafkaMirror Maker Setup Notes

Luckily, Kafka Mirror Maker was used to synchronize data in both locations. Here's your personal notes (not finished). Open JMX Monitoring Kafka's sensitive data using JMX requires opening the JMX port, which can be achieved by setting environment variables.My own practice is to add the following configuration to the kafka-server-start.sh scri ...

Added by cheesemunger on Mon, 29 Jul 2019 04:14:58 +0300

JWT VS Session VS Cookie for ASP.NET Core Web Api

Preface In this article, we will discuss the JWT VS Session. This problem is not too much to think about. When we see that comments and discussions are too heated, we spend a little time to research and summarize. By the way, this is the benefit of blogging. A blog may write with accumulated experience, or it may be learning to share, but it a ...

Added by aznjay on Mon, 29 Jul 2019 02:40:15 +0300

Use of variables based on Tensorflow

Catalog   Code annotations tf.random_normal() tf.Variable() ops.get_collection(ops.GraphKeys.GLOBAL_VARIABLES) tf.variables_initializer Why do variables have to be initialized? Complete code Code annotations tf.random_normal() Used to extract a specified number of values from a value that ...

Added by hermes on Sun, 28 Jul 2019 15:48:56 +0300

Introduction to nginx - load balancing

Load balancing purposes: Forwarding the front-end ultra-high concurrent access to the back-end servers for processing solves the problem that the pressure of a single node is too high, which results in the slow response of Web services, and in serious cases leads to the paralysis of services and the ...

Added by jclarkkent2003 on Sun, 28 Jul 2019 13:48:26 +0300

Cookies, session s of Django components

I. cookie 1.1 Background HTTP protocol is stateless and each request is independent for the server. State can be understood as data generated by the client and server in a session, and statelessness assumes that the data will not be retained. The data generated in the conversation is what we need to preserve, that is to say, to "keep sta ...

Added by Sayian on Fri, 26 Jul 2019 16:56:36 +0300

4 TensorFlow Model Persistence

TensorFlow provides a very simple API to save and restore a neural network model.This API is the tf.train.Saver class.The following code gives a way to save the TesnsorFlow calculation diagram. import tensorflow as tf #Declare two variables and calculate their sum v1 = tf.Variable(tf.constant(1.0, shape = [1]), name = "v1") v2 = tf.Vari ...

Added by Rhiknow on Wed, 24 Jul 2019 21:40:37 +0300

Getting Started with Python Crawlers [7]: Picture Crawl Two of the Hummingbird Web

Picture of Hummingbird Net--Introduction Play fresh today, using a new library aiohttp to speed up our crawls. Installation module routine pip install aiohttp After running, wait, install, want to further study, then the official documents are necessary: https://aiohttp.readthedocs.io/en/stable/ Now you can start writing code. The page we wa ...

Added by sqishy on Wed, 24 Jul 2019 21:14:35 +0300