Analysis of ideas for question D of the 2023 Huashu Cup Mathematical Modeling

0 Question Ideas

(Share the competition questions on CSDN as soon as they come out)
https://blog.csdn.net/dc_sinor

1 Competition information

In order to cultivate students' awareness of innovation and their ability to use mathematical methods and computer technology to solve practical problems, Tianjin Future and Forecasting Science Research Association and China Future Research Association Big Data and Mathematical Model Professional Committee (co-organizer) decided to hold the Wasu Cup National College Student Mathematics Cup Modeling competition. The goal of the competition is to cultivate college students' scientific spirit and ability to use mathematics to solve practical problems, lay a foundation for the cultivation of innovative talents, and cultivate and select outstanding talents for all walks of life.

The competition covers a wide range of fields and may include but is not limited to problems in applied mathematics, statistics, operations research, optimization theory and other related fields. Competition topics are usually derived from practical problems, and participating teams are encouraged to use mathematical modeling methods to solve real-life challenges.

2 Competition time

Registration end time: 12:00 on August 3, 2023

Competition start time: 18:00 on August 3, 2023 (Thursday)

Competition end time: 20:00 on August 6, 2023 (Sunday)

Result announcement time: Estimated to be in mid-to-late August 2023

3 Organizational Structure

organizer:

Wasu Cup National Undergraduate Mathematical Modeling Competition Organizing Committee

Tianjin Future and Forecasting Science Research Society

China Future Research Association Big Data and Mathematical Model Professional Committee (co-organizer)

4 Modeling Common Problem Types

While the competition questions have not been updated yet, Mr. A will give you a summary of the mathematical models commonly used in mathematical modeling of the Wasu Cup. The questions basically belong to the following four types of problems, and the corresponding solutions are also given by Mr. A.

They are:

  • Classification model
  • Optimization model
  • Predictive model
  • Evaluation model

4.1 Classification problem

Discriminant analysis:
Also known as "discrimination method", it is a multi-variable statistical analysis method that determines the type of a research object based on various characteristic values ​​​​under the condition that the classification is determined.
The basic principle is to establish one or more discriminant functions according to certain discriminant criteria; use a large amount of data on the research object to determine the undetermined coefficients in the discriminant function, and calculate the discriminant index; based on this, it is possible to determine what category a certain sample belongs to. When a new sample data is obtained, it is necessary to determine which category of known types the sample belongs to. This type of problem belongs to the discriminant analysis problem.

Cluster analysis:
Cluster analysis or clustering is to divide similar objects into different groups or more subsets through static classification methods, so that member objects in the same subset have similar attributes. Common Including shorter spatial distances in the coordinate system, etc.
Cluster analysis itself is not a specific algorithm, but a general task that needs to be solved. It can be achieved by different algorithms that vary greatly in terms of understanding what clusters are made of and how to find them efficiently.

Neural network classification:
BP neural network is a neural network learning algorithm. It is a hierarchical neural network composed of an input layer, an intermediate layer, and an output layer. The intermediate layer can be expanded to multiple layers. RBF (Radial Basis Function) neural network: The radial basis function (RBF-Radial Basis Function) neural network is a three-layer feedforward network with a single hidden layer. It simulates the structure of neural networks in the human brain that adjust locally and cover each other's receptive fields. Perceptron neural network: It is a neural network with a single layer of computational neurons, and the transfer function of the network is a linear threshold unit. It is mainly used to simulate the perceptual characteristics of the human brain. Linear neural network: It is a relatively simple neural network, consisting of one or more linear neurons. A linear function is used as the transfer function, so the output can be any value. Self-organizing neural network: Self-organizing neural network includes self-organizing competition network, self-organizing feature mapping network, learning vector quantization and other network structure forms. K nearest neighbor algorithm: K nearest neighbor classification algorithm is a theoretically mature method and one of the simplest machine learning algorithms.

4.2 Optimization problem

Linear programming:
mathematical theory and method for studying the extreme value problem of linear objective function under linear constraints. English abbreviation LP. It is an important branch of operations research and is widely used in production planning, logistics and transportation, resource allocation, financial investment and other fields. Modeling method: list the constraints and objective function; draw the feasible region represented by the constraints; find the optimal solution and optimal value of the objective function within the feasible region.

Integer Programming:
The variables in the program (all or part of it) are restricted to integers and are called integer programming. If the variables in a linear model are restricted to integers, it is called integer linear programming. Currently popular methods for solving integer programming are often only applicable to integer linear programming. A type of mathematical programming that requires all or some of the variables in the solution to be integers. From the composition of constraints, it can be subdivided into linear, quadratic and nonlinear integer programming.

Nonlinear programming:
Nonlinear programming is mathematical programming with nonlinear constraints or objective functions. It is an important branch of operations research. Nonlinear programming studies the extreme value problem of an n-element real function under a set of constraints of equality or inequality, and at least one of the objective function and the constraints is a nonlinear function of an unknown quantity. The situation where the objective function and constraints are both linear functions belongs to linear programming.

Dynamic programming:
including knapsack problems, production and operation problems, fund management problems, resource allocation problems, shortest path problems and complex system reliability problems, etc.
Dynamic programming is mainly used to solve optimization problems of dynamic processes divided into stages by time. However, some static programming that has nothing to do with time (such as linear programming and nonlinear programming) can be regarded as a multi-stage decision-making process as long as the time factor is artificially introduced. , can also be easily solved using dynamic programming methods.

Multi-objective programming:
Multi-objective programming is a branch of mathematical programming. Study the optimization of more than one objective function over a given region. Any multi-objective programming problem consists of two basic parts:
(1) more than two objective functions;
(2) several constraints. There are n decision variables, k objective functions, and m constraint equations, then:
Z=F(X) is a k-dimensional function vector, Φ(X) is an m-dimensional function vector; G is an m-dimensional constant vector;

4.3 Prediction problem

Regression fitting prediction
Fitting prediction is the process of establishing a model to approximate the actual data sequence, which is suitable for developmental systems. When building a model, it is usually necessary to specify a meaningful time origin and time unit. Moreover, the model should still make sense as t tends to infinity. The significance of studying fitting prediction as a separate system is to emphasize its "image" nature. The establishment of a prediction model must be as consistent as possible with the actual system. This is the principle of fitting. The degree of fitting can be measured by least squares, maximum likelihood, and minimum absolute deviation.

Gray Forecast
Gray forecast is a prediction made about a gray system. It is a method of predicting systems containing uncertain factors. Gray prediction identifies the degree of dissimilarity in development trends between system factors, that is, performs correlation analysis, and generates and processes original data to find the rules of system changes, generates data sequences with strong regularities, and then establishes corresponding differential equations model to predict the future development trend of things. It constructs a gray prediction model using a series of quantitative values ​​observed at equal intervals that reflect the characteristics of the prediction object, and predicts the characteristic quantity at a certain moment in the future, or the time to reach a certain characteristic quantity.
Markov forecast: It is a method that can be used to forecast the internal human resource supply of an organization. Its basic idea is to find out the rules of personnel changes in the past, so as to predict the trend of future personnel changes. The transformation matrix is ​​actually The transition probability matrix describes the overall form of employee inflow, outflow and internal mobility in the organization, and can be used as the basis for predicting internal labor supply.

BP neural network prediction
BP network (Back-ProPagation Network), also known as back-propagation neural network, continuously corrects the network weights and thresholds through training of sample data to make the error function decrease along the negative gradient direction and approach the expected output. It is a widely used neural network model, mostly used for function approximation, model identification and classification, data compression and time series prediction.

Support vector machine method
Support vector machine (SVM), also known as support vector network [1], is a supervised learning model and related learning algorithms that use classification and regression analysis to analyze data. Given a set of training samples, each training sample is labeled as belonging to one or the other of two categories. The training algorithm of a support vector machine (SVM) creates a model that assigns new samples to one of two categories, making it a non-probabilistic binary linear classifier (although in a probabilistic classification setting, there are things like Prato correction Such methods use support vector machines). The support vector machine model represents samples as mapped points in space, so that samples with a single category can be separated as clearly as possible. All such new samples map to the same space, and their category can be predicted based on which side of the interval they fall on.

4.4 Evaluation issues

The analytic hierarchy process
refers to treating a complex multi-objective decision-making problem as a system, decomposing the goal into multiple goals or criteria, and then decomposing it into several levels of multi-indexes (or criteria, constraints), and calculating the levels through the fuzzy quantification method of qualitative indicators. Single ranking (weight) and total ranking as a systematic method for target (multiple indicators) and multi-scheme optimization decision-making.

The distance method between good and bad solutions
, also known as the ideal solution method, is an effective multi-index evaluation method. This method constructs the positive ideal solution and the negative ideal solution of the evaluation problem, that is, the maximum and minimum values ​​of each indicator, and calculates the relative closeness of each solution to the ideal solution, that is, the distance close to the positive ideal solution and far away from the negative ideal solution. degree, to sort the plans and select the optimal plan.

Fuzzy comprehensive evaluation method
is a comprehensive bid evaluation method based on fuzzy mathematics. This comprehensive evaluation method transforms qualitative evaluation into quantitative evaluation based on the membership theory of fuzzy mathematics, that is, using fuzzy mathematics to make an overall evaluation of things or objects that are restricted by multiple factors. It has the characteristics of clear results and strong systematicness, can better solve fuzzy and difficult-to-quantify problems, and is suitable for solving various non-deterministic problems.

Gray correlation analysis method (gray comprehensive evaluation method)
is a measure of the correlation between factors between two systems that changes with time or different objects, which is called correlation degree. In the process of system development, if the change trends of two factors are consistent, that is, the degree of synchronous change is high, it can be said that the degree of correlation between the two is high; otherwise, the degree of correlation is low. Therefore, the gray correlation analysis method is based on the degree of similarity or dissimilarity in the development trends between factors, that is, the "gray correlation degree", as a method of measuring the degree of correlation between factors.

Canonical correlation analysis method: It is an understanding of the cross-covariance matrix. It is a multivariate statistical analysis method that uses the correlation between pairs of comprehensive variables to reflect the overall correlation between the two groups of indicators. Its basic principle is: in order to grasp the correlation between the two groups of indicators as a whole, two representative comprehensive variables U1 and V1 are extracted from the two groups of variables (respectively, they are the linear coefficients of each variable in the two variable groups). combination), using the correlation between these two comprehensive variables to reflect the overall correlation between the two sets of indicators.

Principal component analysis (dimensionality reduction)
is a statistical method. A set of possibly correlated variables is converted into a set of linearly uncorrelated variables through orthogonal transformation. The converted set of variables is called principal components. When using statistical analysis methods to study multi-variable topics, too many variables will increase the complexity of the topic. People naturally hope that the number of variables is smaller and the information obtained is more. In many cases, there is a certain correlation between variables. When there is a certain correlation between two variables, it can be interpreted that there is a certain overlap in the information reflected by the two variables on this topic. Principal component analysis is to delete the repeated variables (closely related variables) for all the variables originally proposed, and create as few new variables as possible, so that these new variables are not correlated in pairs, and these new variables reflect the The information on the topic is kept as original as possible. A statistical method that tries to recombine the original variables into a new set of several unrelated comprehensive variables, and at the same time, according to actual needs, can take out a few smaller comprehensive variables to reflect as much information as possible about the original variables is called principal component analysis or called principal component analysis. Principal component analysis is also a method used in mathematics to reduce dimensionality.

Factor analysis method (dimensionality reduction)
Factor analysis refers to the study of statistical techniques for extracting common factors from variable groups. It was first proposed by British psychologist CE Spearman. He found that there was a certain correlation between students' scores in various subjects. Students who did well in one subject often had better scores in other subjects, and thus speculated whether there were some potential common factors, or some general intellectual conditions. Affects students' academic performance. Factor analysis can find hidden representative factors among many variables. Grouping variables of the same nature into one factor can reduce the number of variables and test hypotheses about the relationship between variables.

The BP neural network comprehensive evaluation method
is a multi-layer feedforward network trained according to the error back propagation algorithm. It is one of the most widely used neural network models. BP network can learn and store a large number of input-output pattern mapping relationships without revealing the mathematical equations describing this mapping relationship in advance. Its learning rule is to use the steepest descent method and continuously adjust the weights and thresholds of the network through backpropagation to minimize the sum of square errors of the network. The topology structure of BP neural network model includes input layer, hidden layer and output layer.

Guess you like

Origin blog.csdn.net/m0_71450098/article/details/131994788